Align, Disambiguate, and Walk A Unified Approach for Measuring - - PowerPoint PPT Presentation

β–Ά
align disambiguate and walk
SMART_READER_LITE
LIVE PREVIEW

Align, Disambiguate, and Walk A Unified Approach for Measuring - - PowerPoint PPT Presentation

Align, Disambiguate, and Walk A Unified Approach for Measuring Semantic Similarity Semantic Similarity; how similar are a pair of lexical items? Semantic Similarity Semantic Similarity Semantic Similarity Sentence level


slide-1
SLIDE 1

Align, Disambiguate, and Walk

A Unified Approach for Measuring Semantic Similarity

slide-2
SLIDE 2

Semantic Similarity; how similar are a pair of lexical items?

slide-3
SLIDE 3

Semantic Similarity

slide-4
SLIDE 4

Semantic Similarity

slide-5
SLIDE 5

Semantic Similarity

Sentence level

  • (Tsatsaronis et al., 2010)
  • (Kauchak and Barzilay, 2006)
  • (Surdeanu et al., 2011)
  • (Dagan et al., 2006)
slide-6
SLIDE 6

Semantic Similarity

slide-7
SLIDE 7

Semantic Similarity

Word level

  • (Biran et al., 2011)

Locuacious β†’ Talkative

  • Lexical substitution

(McCarthy and Navigli, 2009)

slide-8
SLIDE 8

Semantic Similarity

slide-9
SLIDE 9

Semantic Similarity

Sense level ο€ 

slide-10
SLIDE 10

Semantic Similarity

Sense level

  • (Snow et al., 2007)
  • (Neely et al., 1989)
slide-11
SLIDE 11

Exisiting Similarity Measures

slide-12
SLIDE 12

Exisiting Similarity Measures

Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-13
SLIDE 13

Exisiting Similarity Measures

Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-14
SLIDE 14

Exisiting Similarity Measures

Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-15
SLIDE 15

Exisiting Similarity Measures

Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-16
SLIDE 16

Exisiting Similarity Measures

Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-17
SLIDE 17

Exisiting Similarity Measures

Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-18
SLIDE 18

Exisiting Similarity Measures

Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)

slide-19
SLIDE 19

Contribution

slide-20
SLIDE 20

Contribution

slide-21
SLIDE 21

Contribution

slide-22
SLIDE 22

Contribution

slide-23
SLIDE 23

Advantage 1

Unified representation

slide-24
SLIDE 24

Advantage 2

Cross-level semantic similarity

𝑀𝑑. 𝑀𝑑.

slide-25
SLIDE 25

Advantage 3

Sense-level operation

slide-26
SLIDE 26

Outline

slide-27
SLIDE 27

How Does it work?

slide-28
SLIDE 28

How Does it work?

slide-29
SLIDE 29

How Does it work?

slide-30
SLIDE 30

Semantic Signature

slide-31
SLIDE 31

Semantic Signature

slide-32
SLIDE 32

Semantic Signature

slide-33
SLIDE 33

Semantic Signature

slide-34
SLIDE 34

Semantic Signature

slide-35
SLIDE 35

Semantic Signature

slide-36
SLIDE 36

Semantic Signature

slide-37
SLIDE 37

Semantic Signature

slide-38
SLIDE 38

Semantic Signature

slide-39
SLIDE 39

Semantic Signature

slide-40
SLIDE 40

Semantic Signature

slide-41
SLIDE 41

Semantic Signature

  • ver all synsets in WordNet

. . .

slide-42
SLIDE 42

Semantic Signature

  • ver all synsets in WordNet

. . .

slide-43
SLIDE 43

Semantic Signature

  • ver all synsets in WordNet

. . .

slide-44
SLIDE 44

Semantic Signature

slide-45
SLIDE 45

Semantic Signature

slide-46
SLIDE 46

Semantic Signature

slide-47
SLIDE 47

Semantic Signature

slide-48
SLIDE 48

Semantic Signature

slide-49
SLIDE 49

Semantic Signature

slide-50
SLIDE 50

Personalized PageRank

slide-51
SLIDE 51

Personalized PageRank

slide-52
SLIDE 52

Personalized PageRank

slide-53
SLIDE 53

Personalized PageRank

n

woman 1

n

food 1

v

fry 2

slide-54
SLIDE 54

Personalized PageRank

n

woman 1

n

food 1

v

fry 2

slide-55
SLIDE 55

Personalized PageRank

v

fry 2

n

woman 1

n

food 1

v

fry 2

slide-56
SLIDE 56

Personalized PageRank

v

fry 2

n

food 1

n

woman 1

n

food 1

v

fry 2

slide-57
SLIDE 57

Personalized PageRank

v

fry 2

n

food 1

n

woman 1

n

woman 1

n

food 1

v

fry 2

slide-58
SLIDE 58

Personalized PageRank

v

fry 2

n

food 1

n

woman 1

n

woman 1

n

food 1

v

fry 2

slide-59
SLIDE 59

v

fry

2

n

food

1

n

cooking 1

v

cook 3

n

fat 1

n

french_fries 1

n

dish 2

n

nutriment 1

n

food

2

n

beverage

1

slide-60
SLIDE 60

v

fry

2

n

food

1

n

cooking 1

v

cook 3

n

fat 1

n

french_fries 1

n

dish 2

n

nutriment 1

n

food

2

n

beverage

1

slide-61
SLIDE 61

Comparing Semantic Signatures

slide-62
SLIDE 62

Comparing Semantic Signatures

  • – Cosine
  • – Weighted Overlap

– Top-k Jaccard

slide-63
SLIDE 63

Comparing Semantic Signatures

Weighted Overlap

slide-64
SLIDE 64

Comparing Semantic Signatures

Weighted Overlap

slide-65
SLIDE 65

Comparing Semantic Signatures

Weighted Overlap

slide-66
SLIDE 66

Comparing Semantic Signatures

Weighted Overlap

slide-67
SLIDE 67

Comparing Semantic Signatures

Weighted Overlap

slide-68
SLIDE 68

Comparing Semantic Signatures

T

  • p-𝑙 Jaccard
slide-69
SLIDE 69

Comparing Semantic Signatures

T

  • p-𝑙 Jaccard

𝑙 = 4

slide-70
SLIDE 70

Comparing Semantic Signatures

T

  • p-𝑙 Jaccard

𝑙 = 4

slide-71
SLIDE 71

Comparing Semantic Signatures

T

  • p-𝑙 Jaccard

𝑙 = 4

slide-72
SLIDE 72

Alignment-based disambiguation

slide-73
SLIDE 73

Alignment-based disambiguation

slide-74
SLIDE 74

Why is disambiguation needed?

slide-75
SLIDE 75

Why is disambiguation needed?

slide-76
SLIDE 76

Why is disambiguation needed?

slide-77
SLIDE 77

Why is disambiguation needed?

slide-78
SLIDE 78

Why is disambiguation needed?

slide-79
SLIDE 79

Alignment-based disambiguation

slide-80
SLIDE 80

Alignment-based disambiguation

slide-81
SLIDE 81

Alignment-based disambiguation

slide-82
SLIDE 82

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

fire worker fire fire worker

. . .

v

1

v

2

v

3

n

1

n

2

. . .

n

Alignment-based disambiguation

slide-83
SLIDE 83

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

fire worker fire fire worker

. . .

v

1

v

2

v

3

n

1

n

2

. . .

n

Alignment-based disambiguation

slide-84
SLIDE 84

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

n

Alignment-based disambiguation

slide-85
SLIDE 85

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

n

Alignment-based disambiguation

slide-86
SLIDE 86

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

n

Alignment-based disambiguation

0.5

slide-87
SLIDE 87

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

n

Alignment-based disambiguation

0.5 Tversky (1977) Markman and Gentner (1993)

slide-88
SLIDE 88

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

n

Alignment-based disambiguation

0.5 0.3 Tversky (1977) Markman and Gentner (1993)

slide-89
SLIDE 89

terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

Alignment-based disambiguation

0.3

n

manager 2

n

manager

n

1

0.5 0.3

slide-90
SLIDE 90

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

fire worker fire fire worker

. . .

v

1

v

2

v

3

n

1

n

2

. . .

n

fire v

4

Alignment-based disambiguation

slide-91
SLIDE 91

manager terminate work boss boss work

. . .

terminate terminate terminate employee work

. . .

n

manager 2

n

1

n

1

v

1

v

2

v

3

v

4

n

2

n

1

n

3

n

1

n

2

fire worker fire fire worker

. . .

v

1

v

2

v

3

n

1

n

2

. . .

n

fire v

4

Alignment-based disambiguation

slide-92
SLIDE 92

Outline

slide-93
SLIDE 93

Experiments

  • – Semantic Textual Similarity (SemEval-2012)
slide-94
SLIDE 94

Experiments

  • – Semantic Textual Similarity (SemEval-2012)
  • – Synonymy recognition (TOEFL dataset)

– Correlation-based (RG-65 dataset)

slide-95
SLIDE 95

Experiments

  • – Semantic Textual Similarity (SemEval-2012)
  • – Synonymy recognition (TOEFL dataset)

– Correlation-based (RG-65 dataset)

  • – Coarsening WordNet sense inventory
slide-96
SLIDE 96

Experiment 1

Similarity at Sentence level

  • – 5 datasets

– Three evaluation measures

  • ALL, ALLnrm, and Mean
slide-97
SLIDE 97

Experiment 1

Similarity at Sentence level

  • – 5 datasets

– Three evaluation measures

  • ALL, ALLnrm, and Mean
  • – UKP2 (BΓ€r et al., 2012)

– TLSim and TLSyn (Ε ariΔ‡ et al., 2012)

slide-98
SLIDE 98

Experiment 1

Similarity at Sentence level

Features

– Main features

  • Cosine
  • Weighted Overlap
  • Top-k Jaccard
slide-99
SLIDE 99

Experiment 1

Similarity at Sentence level

Features

– Main features

  • Cosine
  • Weighted Overlap
  • Top-k Jaccard

– String-based features

  • Longest common substring
  • Longest common subsequence
  • Greedy string tiling
  • Character/word n-grams
slide-100
SLIDE 100

STS Results

Experiment 1

Similarity at Sentence level

ADW ADW ADW UKP2 UKP2 UKP2 TLsyn TLsyn TLsyn TLsim TLsim TLsim

slide-101
SLIDE 101

STS Results

Experiment 1

Similarity at Sentence level

ADW ADW ADW UKP2 UKP2 UKP2 TLsyn TLsyn TLsyn TLsim TLsim TLsim

slide-102
SLIDE 102

Experiments

  • – Semantic Textual Similarity (Semeval-12)
  • – Synonymy recognition (TOEFL dataset)

– Correlation-based (RG-65 dataset)

  • – Coarsening WordNet sense inventory
slide-103
SLIDE 103

Experiment 2

Similarity at Word Level

slide-104
SLIDE 104

Experiment 2

Similarity at Word Level

slide-105
SLIDE 105

Experiment 2

Similarity at Word Level

slide-106
SLIDE 106

Similarity at Word Level

Synonym Recognition

  • TOEFL dataset (Landauer and Dumais, 1997)

– 80 multiple choice questions – Human test takers: 64.5% only

slide-107
SLIDE 107

Similarity at Word Level

Synonym Recognition

Accuracy on TOEFL dataset

75 80 85 90 95 100

slide-108
SLIDE 108

Similarity at Word Level

Judgment Correlation

  • Dataset: RG-65 (Rubenstein and Goodenough, 1965)

– 65 word pairs

  • judged by 51 human subjects

– Scale of 0 β†’ 4

slide-109
SLIDE 109

Similarity at Word Level

Judgment Correlation

Spearman correlation, RG-65 dataset

slide-110
SLIDE 110

Similarity at Word Level

Judgment Correlation

Spearman correlation, RG-65 dataset

slide-111
SLIDE 111

Experiments

  • – Semantic Textual Similarity (Semeval-12)
  • – Synonymy recognition (TOEFL dataset)

– Correlation-based (RG-65 dataset)

  • – Coarsening WordNet sense inventory
slide-112
SLIDE 112

Experiment 3

Similarity at Sense Level

  • Coarse-graining WordNet
slide-113
SLIDE 113

Experiment 3

Similarity at Sense Level

  • Coarse-graining WordNet

Navigli (2006)

slide-114
SLIDE 114

Experiment 3

Similarity at Sense Level

  • Coarse-graining WordNet

Snow et al. (2007) Navigli (2006)

slide-115
SLIDE 115

Experiment 3

Similarity at Sense Level

  • Binary classification: Merged or not-merged
slide-116
SLIDE 116

Experiment 3

Similarity at Sense Level

  • Binary classification: Merged or not-merged

(Kilgarriff, 2001) (Hovy et al., 2006)

about 3500 sense pairs (Noun) about 5000 sense pairs (Verb) about 16000 sense pairs (Noun) about 31000 sense pairs (Verb)

slide-117
SLIDE 117

Experiment 3

Similarity at Sense Level

  • Binary classification: Merge or not-merged

if similarity β‰₯ 𝑒 π‘π‘’β„Žπ‘“π‘ π‘₯𝑗𝑑𝑓

slide-118
SLIDE 118

Experiment 3

Similarity at Sense Level

  • Binary classification: Merge or not-merged

if similarity β‰₯ 𝑒 π‘π‘’β„Žπ‘“π‘ π‘₯𝑗𝑑𝑓

slide-119
SLIDE 119

Experiment 3

Similarity at Sense Level

F-score on OntoNotes dataset

0.41 0.42 0.42 0.37 0.22

Noun

0.52 0.54 0.53 0.45 0.37

Verb

slide-120
SLIDE 120

Experiment 3

Similarity at Sense Level

F-score on Senseval-2 dataset

0.44 0.47 0.47 0.42 0.37

Noun

0.49 0.5 0.49 0.43 0.29

Verb

slide-121
SLIDE 121

Conclusions

  • A unified approach for computing semantic

similarity for any pair of lexical items

  • Experiments with SOA performance

– Sense level (sense coarsening) – Word level (synonymy recognition and judgment) – Sentence level (Semantic Textual Similarity)

slide-122
SLIDE 122

Future Direction

  • Larger sense inventories (e.g., BabelNet)
slide-123
SLIDE 123

Future Direction

  • Larger sense inventories (e.g., BabelNet)
  • Cross-level semantic similarity
slide-124
SLIDE 124

Future Direction

  • Larger sense inventories (e.g., BabelNet)
  • Cross-level semantic similarity

Create datasets for cross-level similarity

– Future Semeval task?

slide-125
SLIDE 125

n

listening 1

n

thank_you 1

n

gratitude 1

n

thanks 1

n

expression 3

v

heed 1

n

sensing 2

v

listen 2

n

auscultation 1

n

ear 2

n

hearing 6

j

audio-lingual 1

v

listen 1

slide-126
SLIDE 126
slide-127
SLIDE 127

STS-13

System HDL OnWN FNWN SMT mean Rank Dkpro 0.735 0.735 0.341 0.323 0.565 6 TakeLab 0.486 0.633 0.269 0.279 0.434 58 ADW (STS-13) 0.621 0.511 0.446 0.384 0.502 34 ADW (All) GP 0.717 0.697 0.411 0.272 0.538 20 ADW (All) LR 0.667 0.735 0.409 0.374 0.565 6