Align, Disambiguate, and Walk A Unified Approach for Measuring - - PowerPoint PPT Presentation
Align, Disambiguate, and Walk A Unified Approach for Measuring - - PowerPoint PPT Presentation
Align, Disambiguate, and Walk A Unified Approach for Measuring Semantic Similarity Semantic Similarity; how similar are a pair of lexical items? Semantic Similarity Semantic Similarity Semantic Similarity Sentence level
Semantic Similarity; how similar are a pair of lexical items?
Semantic Similarity
Semantic Similarity
Semantic Similarity
Sentence level
- (Tsatsaronis et al., 2010)
- (Kauchak and Barzilay, 2006)
- (Surdeanu et al., 2011)
- (Dagan et al., 2006)
Semantic Similarity
Semantic Similarity
Word level
- (Biran et al., 2011)
Locuacious β Talkative
- Lexical substitution
(McCarthy and Navigli, 2009)
Semantic Similarity
Semantic Similarity
Sense level ο
Semantic Similarity
Sense level
- (Snow et al., 2007)
- (Neely et al., 1989)
Exisiting Similarity Measures
Exisiting Similarity Measures
Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Exisiting Similarity Measures
Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Exisiting Similarity Measures
Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Exisiting Similarity Measures
Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Exisiting Similarity Measures
Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Exisiting Similarity Measures
Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Exisiting Similarity Measures
Patwardan (2003) Banerjee and Pederson (2003) Hirst and St-Onge (1998) Lin (1998) Jiang and Conrath (1997) Resnik (1995) Sussna (1993, 1997) Wu and Palmer (1994) Leacock and Chodorow (1998) Salton and McGill (1983) Gabrilovich and Markovitch (2007) Radinsky et al. (2011) Ramage et al. (2009) Yeh et al., (2009) Turney (2007) Landauer et al. (1998) Allison and Dix (1986) Gusfield (1997) Wise (1996) Keselj et al. (2003)
Contribution
Contribution
Contribution
Contribution
Advantage 1
Unified representation
Advantage 2
Cross-level semantic similarity
π€π‘. π€π‘.
Advantage 3
Sense-level operation
Outline
How Does it work?
How Does it work?
How Does it work?
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
- ver all synsets in WordNet
. . .
Semantic Signature
- ver all synsets in WordNet
. . .
Semantic Signature
- ver all synsets in WordNet
. . .
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Semantic Signature
Personalized PageRank
Personalized PageRank
Personalized PageRank
Personalized PageRank
n
woman 1
n
food 1
v
fry 2
Personalized PageRank
n
woman 1
n
food 1
v
fry 2
Personalized PageRank
v
fry 2
n
woman 1
n
food 1
v
fry 2
Personalized PageRank
v
fry 2
n
food 1
n
woman 1
n
food 1
v
fry 2
Personalized PageRank
v
fry 2
n
food 1
n
woman 1
n
woman 1
n
food 1
v
fry 2
Personalized PageRank
v
fry 2
n
food 1
n
woman 1
n
woman 1
n
food 1
v
fry 2
v
fry
2
n
food
1
n
cooking 1
v
cook 3
n
fat 1
n
french_fries 1
n
dish 2
n
nutriment 1
n
food
2
n
beverage
1
v
fry
2
n
food
1
n
cooking 1
v
cook 3
n
fat 1
n
french_fries 1
n
dish 2
n
nutriment 1
n
food
2
n
beverage
1
Comparing Semantic Signatures
Comparing Semantic Signatures
- β Cosine
- β Weighted Overlap
β Top-k Jaccard
Comparing Semantic Signatures
Weighted Overlap
Comparing Semantic Signatures
Weighted Overlap
Comparing Semantic Signatures
Weighted Overlap
Comparing Semantic Signatures
Weighted Overlap
Comparing Semantic Signatures
Weighted Overlap
Comparing Semantic Signatures
T
- p-π Jaccard
Comparing Semantic Signatures
T
- p-π Jaccard
π = 4
Comparing Semantic Signatures
T
- p-π Jaccard
π = 4
Comparing Semantic Signatures
T
- p-π Jaccard
π = 4
Alignment-based disambiguation
Alignment-based disambiguation
Why is disambiguation needed?
Why is disambiguation needed?
Why is disambiguation needed?
Why is disambiguation needed?
Why is disambiguation needed?
Alignment-based disambiguation
Alignment-based disambiguation
Alignment-based disambiguation
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
fire worker fire fire worker
. . .
v
1
v
2
v
3
n
1
n
2
. . .
n
Alignment-based disambiguation
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
fire worker fire fire worker
. . .
v
1
v
2
v
3
n
1
n
2
. . .
n
Alignment-based disambiguation
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
n
Alignment-based disambiguation
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
n
Alignment-based disambiguation
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
n
Alignment-based disambiguation
0.5
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
n
Alignment-based disambiguation
0.5 Tversky (1977) Markman and Gentner (1993)
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
n
Alignment-based disambiguation
0.5 0.3 Tversky (1977) Markman and Gentner (1993)
terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
Alignment-based disambiguation
0.3
n
manager 2
n
manager
n
1
0.5 0.3
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
fire worker fire fire worker
. . .
v
1
v
2
v
3
n
1
n
2
. . .
n
fire v
4
Alignment-based disambiguation
manager terminate work boss boss work
. . .
terminate terminate terminate employee work
. . .
n
manager 2
n
1
n
1
v
1
v
2
v
3
v
4
n
2
n
1
n
3
n
1
n
2
fire worker fire fire worker
. . .
v
1
v
2
v
3
n
1
n
2
. . .
n
fire v
4
Alignment-based disambiguation
Outline
Experiments
- β Semantic Textual Similarity (SemEval-2012)
Experiments
- β Semantic Textual Similarity (SemEval-2012)
- β Synonymy recognition (TOEFL dataset)
β Correlation-based (RG-65 dataset)
Experiments
- β Semantic Textual Similarity (SemEval-2012)
- β Synonymy recognition (TOEFL dataset)
β Correlation-based (RG-65 dataset)
- β Coarsening WordNet sense inventory
Experiment 1
Similarity at Sentence level
- β 5 datasets
β Three evaluation measures
- ALL, ALLnrm, and Mean
Experiment 1
Similarity at Sentence level
- β 5 datasets
β Three evaluation measures
- ALL, ALLnrm, and Mean
- β UKP2 (BΓ€r et al., 2012)
β TLSim and TLSyn (Ε ariΔ et al., 2012)
Experiment 1
Similarity at Sentence level
Features
β Main features
- Cosine
- Weighted Overlap
- Top-k Jaccard
Experiment 1
Similarity at Sentence level
Features
β Main features
- Cosine
- Weighted Overlap
- Top-k Jaccard
β String-based features
- Longest common substring
- Longest common subsequence
- Greedy string tiling
- Character/word n-grams
STS Results
Experiment 1
Similarity at Sentence level
ADW ADW ADW UKP2 UKP2 UKP2 TLsyn TLsyn TLsyn TLsim TLsim TLsim
STS Results
Experiment 1
Similarity at Sentence level
ADW ADW ADW UKP2 UKP2 UKP2 TLsyn TLsyn TLsyn TLsim TLsim TLsim
Experiments
- β Semantic Textual Similarity (Semeval-12)
- β Synonymy recognition (TOEFL dataset)
β Correlation-based (RG-65 dataset)
- β Coarsening WordNet sense inventory
Experiment 2
Similarity at Word Level
Experiment 2
Similarity at Word Level
Experiment 2
Similarity at Word Level
Similarity at Word Level
Synonym Recognition
- TOEFL dataset (Landauer and Dumais, 1997)
β 80 multiple choice questions β Human test takers: 64.5% only
Similarity at Word Level
Synonym Recognition
Accuracy on TOEFL dataset
75 80 85 90 95 100
Similarity at Word Level
Judgment Correlation
- Dataset: RG-65 (Rubenstein and Goodenough, 1965)
β 65 word pairs
- judged by 51 human subjects
β Scale of 0 β 4
Similarity at Word Level
Judgment Correlation
Spearman correlation, RG-65 dataset
Similarity at Word Level
Judgment Correlation
Spearman correlation, RG-65 dataset
Experiments
- β Semantic Textual Similarity (Semeval-12)
- β Synonymy recognition (TOEFL dataset)
β Correlation-based (RG-65 dataset)
- β Coarsening WordNet sense inventory
Experiment 3
Similarity at Sense Level
- Coarse-graining WordNet
Experiment 3
Similarity at Sense Level
- Coarse-graining WordNet
Navigli (2006)
Experiment 3
Similarity at Sense Level
- Coarse-graining WordNet
Snow et al. (2007) Navigli (2006)
Experiment 3
Similarity at Sense Level
- Binary classification: Merged or not-merged
Experiment 3
Similarity at Sense Level
- Binary classification: Merged or not-merged
(Kilgarriff, 2001) (Hovy et al., 2006)
about 3500 sense pairs (Noun) about 5000 sense pairs (Verb) about 16000 sense pairs (Noun) about 31000 sense pairs (Verb)
Experiment 3
Similarity at Sense Level
- Binary classification: Merge or not-merged
if similarity β₯ π’ ππ’βππ π₯ππ‘π
Experiment 3
Similarity at Sense Level
- Binary classification: Merge or not-merged
if similarity β₯ π’ ππ’βππ π₯ππ‘π
Experiment 3
Similarity at Sense Level
F-score on OntoNotes dataset
0.41 0.42 0.42 0.37 0.22
Noun
0.52 0.54 0.53 0.45 0.37
Verb
Experiment 3
Similarity at Sense Level
F-score on Senseval-2 dataset
0.44 0.47 0.47 0.42 0.37
Noun
0.49 0.5 0.49 0.43 0.29
Verb
Conclusions
- A unified approach for computing semantic
similarity for any pair of lexical items
- Experiments with SOA performance
β Sense level (sense coarsening) β Word level (synonymy recognition and judgment) β Sentence level (Semantic Textual Similarity)
Future Direction
- Larger sense inventories (e.g., BabelNet)
Future Direction
- Larger sense inventories (e.g., BabelNet)
- Cross-level semantic similarity
Future Direction
- Larger sense inventories (e.g., BabelNet)
- Cross-level semantic similarity
Create datasets for cross-level similarity
β Future Semeval task?
n
listening 1
n
thank_you 1
n
gratitude 1
n
thanks 1
n
expression 3
v
heed 1
n
sensing 2
v
listen 2
n
auscultation 1
n
ear 2
n
hearing 6
j
audio-lingual 1
v
listen 1
STS-13
System HDL OnWN FNWN SMT mean Rank Dkpro 0.735 0.735 0.341 0.323 0.565 6 TakeLab 0.486 0.633 0.269 0.279 0.434 58 ADW (STS-13) 0.621 0.511 0.446 0.384 0.502 34 ADW (All) GP 0.717 0.697 0.411 0.272 0.538 20 ADW (All) LR 0.667 0.735 0.409 0.374 0.565 6