word sense disambiguation
play

Word Sense Disambiguation Ling571 Deep Processing Techniques for - PowerPoint PPT Presentation

Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP March 3, 2014 Distributional Similarity Questions What is the right neighborhood? What is the context? How should we weight the features? How can we


  1. Word Sense Disambiguation Ling571 Deep Processing Techniques for NLP March 3, 2014

  2. Distributional Similarity Questions — What is the right neighborhood? — What is the context? — How should we weight the features? — How can we compute similarity between vectors?

  3. Feature Vector Design — Window size: — How many words in the neighborhood? — Tradeoff: — +/- 500 words: ‘topical context’ — +/- 1 or 2 words: collocations, predicate-argument — Only words in some grammatical relation — Parse text (dependency) — Include subj-verb; verb-obj; adj-mod — NxR vector: word x relation

  4. Example Lin Relation Vector

  5. Weighting Features — Baseline: Binary (0/1) — Minimally informative — Can’t capture intuition that frequent features informative — Frequency or Probability: P ( f | w ) = count ( f , w ) count ( w ) — Better but, — Can overweight a priori frequent features — Chance cooccurrence

  6. Pointwise Mutual Information P ( w , f ) assoc PMI ( w , f ) = log 2 P ( w ) P ( f ) PMI: - Contrasts observed cooccurrence - With that expected by chance (if independent) - Generally only use positive values - Negatives inaccurate unless corpus huge

  7. Vector Similarity — Euclidean or Manhattan distances: — Too sensitive to extreme values sim dot − product (  v ,  w ) =  v •  N — Dot product: ∑ w = v i × w i — Favors long vectors: i = 1 — More features or higher values N ∑ v i × w i sim cos ine (  v ,  — Cosine: w ) = i = 1 N N ∑ 2 ∑ 2 v i w i i = 1 i = 1

  8. Distributional Similarity for Word Sense Disambiguation

  9. Schutze ’ s Word Space — Build a co-occurrence matrix — Restrict Vocabulary to 4 letter sequences — Similar effect to stemming — Exclude Very Frequent - Articles, Affixes — Entries in 5000-5000 Matrix — Apply Singular Value Decomposition (SVD) — Reduce to 97 dimensions — Word Context — 4grams within 1001 Characters — Sum & Normalize Vectors for each 4gram — Distances between Vectors by dot product

  10. Schutze ’ s Word Space — Word Sense Disambiguation — Context Vectors of All Instances of Word — Automatically Cluster Context Vectors — Hand-label Clusters with Sense Tag — Tag New Instance with Nearest Cluster

  11. There are more kinds of plants and animals in the rainforests than anywhere else on Earth. Over half of the millions of known species of plants and animals live in the rainforest. Many are found nowhere else. There are even plants and animals in the rainforest that we have not yet discovered. Biological Example The Paulus company was founded in 1938. Since those days the product range has been the subject of constant expansions and is brought up continuously to correspond with the state of the art. We ’ re engineering, manufacturing and commissioning world- wide ready-to-run plants packed with our comprehensive know- how. Our Product Range includes pneumatic conveying systems for carbon, carbide, sand, lime andmany others. We use reagent injection in molten metal for the… Industrial Example Label the First Use of “ Plant ”

  12. Sense Selection in “ Word Space ” — Build a Context Vector — 1,001 character window - Whole Article — Compare Vector Distances to Sense Clusters — Only 3 Content Words in Common — Distant Context Vectors — Clusters - Build Automatically, Label Manually — Result: 2 Different, Correct Senses — 92% on Pair-wise tasks

  13. Odd Cluster Examples — The “ Ste. ” Cluster: — Dry Oyster Whisky Hot Float Ice

  14. Odd Cluster Examples — The “ Ste. ” Cluster: — Dry Oyster Whisky Hot Float Ice — Why? – River name

  15. Odd Cluster Examples — The “ Ste. ” Cluster: — Dry Oyster Whisky Hot Float Ice — Why? – River name — Learning the Corpus, not the Sense — Keeping cluster: — Bring Hoping Wiping Could Should Some Them Rest

  16. Taxonomy of Contextual Information — Topical Content — Word Associations — Syntactic Constraints — Selectional Preferences — World Knowledge & Inference

  17. The Question of Context — Shared Intuition: — Context -> Sense — Area of Disagreement: — What is context? — Wide vs Narrow Window — Word Co-occurrences — Best model, best weighting — Still active focus of research

  18. Minimally Supervised WSD — Yarowsky’s algorithm (1995) — Bootstrapping approach: — Use small labeled seedset to iteratively train — Builds on 2 key insights: — One Sense Per Discourse — Word appearing multiple times in text has same sense — Corpus of 37232 bass instances: always single sense — One Sense Per Collocation — Local phrases select single sense — Fish -> Bass 1 — Play -> Bass 2

  19. Yarowsky ’ s Algorithm — Training Decision Lists — 1. Pick Seed Instances & Tag — 2. Find Collocations: Word Left, Word Right, Word +K — (A) Calculate Informativeness on Tagged Set, — Order: abs (log P ( Sense 1 | Collocation ) P ( Sense 2 | Collocation )) — (B) Tag New Instances with Rules — (C) Apply 1 Sense/Discourse — (D) If Still Unlabeled, Go To 2 — 3. Apply 1 Sense/Discourse — Disambiguation: First Rule Matched

  20. Yarowsky Decision List

  21. Iterative Updating

  22. There are more kinds of plants and animals in the rainforests than anywhere else on Earth. Over half of the millions of known species of plants and animals live in the rainforest. Many are found nowhere else. There are even plants and animals in the rainforest that we have not yet discovered. Biological Example The Paulus company was founded in 1938. Since those days the product range has been the subject of constant expansions and is brought up continuously to correspond with the state of the art. We ’ re engineering, manufacturing and commissioning world- wide ready-to-run plants packed with our comprehensive know- how. Our Product Range includes pneumatic conveying systems for carbon, carbide, sand, lime andmany others. We use reagent injection in molten metal for the… Industrial Example Label the First Use of “ Plant ”

  23. Sense Choice With Collocational Decision Lists — Create Initial Decision List abs (log P ( Sense 1 | Collocation ) — Rules Ordered by P ( Sense 2 | Collocation )) — Check nearby Word Groups (Collocations) — Biology: “ Animal ” in + 2-10 words — Industry: “ Manufacturing ” in + 2-10 words — Result: Correct Selection — 95% on Pair-wise tasks

  24. Naïve Bayes ’ Approach — Supervised learning approach — Input: feature vector X label — Best sense = most probable sense given f  ˆ s = argmax P ( s | f ) s ∈ S  P ( f | s ) P ( s )  ˆ s = argmax P ( f ) s ∈ S

  25. Naïve Bayes ’ Approach — Issue: — Data sparseness: full feature vector rarely seen — “ Naïve ” assumption: — Features independent given sense  n ∏ P ( f | s ) ≈ P ( f j | s ) j = 1 n ∏ ˆ s = argmax P ( s ) P ( f j | s ) s ∈ S j = 1

  26. Training NB Classifier n ∏ ˆ s = argmax P ( s ) P ( f j | s ) s ∈ S j = 1 — Estimate P(s): — Prior P ( s i ) = count ( s i , w j ) count ( w j ) P ( f j | s ) = count ( f j , s ) — Estimate P(f j |s) count ( s ) — Issues: — Underflow => log prob — Sparseness => smoothing

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend