collins parsing
play

Collins Parsing Victor, Ydng Zhu Outline Introduction Basic Model - PowerPoint PPT Presentation

Collins Parsing Victor, Ydng Zhu Outline Introduction Basic Model Representation Calculation Three generative models Models Practice issues Evaluation 2 Introduction Michael Collins PhD Thesis, 1999


  1. Collins Parsing Victor, Yùdōng Zhōu

  2. Outline • Introduction • Basic Model • Representation • Calculation • Three generative models • Models • Practice issues • Evaluation 2

  3. Introduction • Michael Collins PhD Thesis, 1999 • Head-Driven (Lexical info) • Statistical, supervised • Input: Tagged sentence • Output: phrase-structure tree 3

  4. Basic Model • Task: Given a sentence, candidate trees and probabilities, find best parsing tree • In this model: T=(B,D), where • B=set of baseNPs • D=set of dependencies 4

  5. Basic Model An Example 5

  6. Basic Model • Dependency Set • Step one: Find head child • Eg. S  <NP VP> • Step two: Extract head modifier • Eg. NP modify VP, with rule S  <NP VP> 6

  7. Basic Model Notation: AF(j) = (hj, Rj) Eg. AF(1) = (5, <NP,S,VP>). where w1=Smith, w5=announced D = {(AF(1),AF(2)...AF(m)} P(T | S)=P(B,D | S)= P(B | S)* P(D | S,B) 7

  8. Basic Model • Calculation • Dependency Probability: Training data 8

  9. Basic Model • Calculation • Dependency Probability: Training data • Distance Measure Eg . Second “of” will reduce the probability: Shaw, based in Dalton, Ga., has annual sales of about $1.18 billion, and has economies of scale and lower raw-material costs that are expected to boost the profitability of Armstrong's brands, sold under the Armstrong and Evans-Black names . 9

  10. Basic Model • Calculation • Dependency Probability: Training data • Distance Measure • Sparse Data • Solved by smoothing 10

  11. Three Generative Models Generative Model Discrimitive Model joint probability conditional distribution distribution -  P(T,S) -  P(T|S) 11

  12. Modal 1 Representation: 12

  13. Modal 1 • Calculation: PCFG based • Generate head, P H (H|P,h) • Generate right modifier, P R (R i (r i )|P,h,H) • Until STOP symbol, R m+1 (r m+1 ) • Generate left modifier, P l (L i (l i )|P,h,H) • Example: • S(bought)  NP(week) NP(Marks) VP(bought) • P h (VP|S,bought) *P l (NP(Marks)|S,VP,bought) *P l (NP(week)|S,VP,bought) *P l (STOP|S,VP,bought) 13 *P r (STOP|S,VP,bought)

  14. Modal 1 • Calculation • Distance Measure • P R (R i (r i )|P,h,H, R 1 (r 1 ),…R i-1 (r i-1 ) ) =P R (R i (r i )|P,h,H) In Previous Formula =P R (R i (r i )|P,h,H, distance r (i-1) ) 14

  15. Modal 2 • Complement/Adjunct distinction • Reasons doing this while parsing: • Lexical info/additional knowledge needed 15 • Help parsing accuracy

  16. Modal 2 • Identifying Complement in Penn Treebank • Rule based • One incorrect Example: • How to get the correct one? 16

  17. Modal 2 • Subcategorisation Frames • Generate head, P H (H|P,h) • Generate left and right subcat frames, LC and RC, P lc (LC|P,H,h) and P rc (RC|P,H,h) • Generate right modifier, (and then left modifier) P R (R i (r i )|P,h,H, distance r (i-1), RC ) …… 17

  18. Modal 2 • Subcat Frames: Example • S(bought)  NP(week) NP-C(Marks) VP(bought) • P h (VP|S,bought) * P lc ({NP-C}|S,VP,bought) * P rc ({ }|S,VP,bought) * P l (NP-C(Marks)|S,VP,bought, {NP-C} )* P l (NP(week)|S,VP,bought, { } ) * P l (STOP|S,VP,bought, { }) * P r (STOP|S,VP,bought, { }) P lc ({NP-C,NP-C}|S,VP,bought) will be quite small Thus achieve the correct parse 18

  19. Modal 3 • Traces and Wh-movement • Example 1 The store (SBAR which TRACE bought Brooks Brothers) • Example 2 The store (SBAR which Marks bought TRACE) • Example 3 The store (SBAR which Marks bought Brooks Brothers from TRACE) 19

  20. Modal 3 • +gap feature added • Introduce parameter P G (G|P,h,H) 20 where G is Head, Left or Right

  21. Practice Issues • Smoothing • Eg. P H estimation e 2 =P H (H|P,t) • Final estimation: • Unknown words 21

  22. Evaluation • Training data: • Section 02-21, Wall Street Journal portion • (Approximately 40,000 sentences) • Testing data: • Section 23 (2,416 sentences) 22

  23. Evaluation • PARSEVAL measures • Label Precision = number of correct constituents in proposed parse number of constituents in proposed parse • Label Recall = number of correct constituents in proposed parse number of constituents in treebank parse • Crossing Brackets = number of constituents which violate constituent boundaries with a constituent in the treebank parse. 23

  24. Evaluation • Collins 96 vs. Model 1 • Model 1 better on unary rules and distance measures • Model 2 vs. Model 3 • For 436 trace cases in testing data, Model 3 has 24 precision/recall 93.3%/90.1%

  25. Q&A 25

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend