Entity Clustering Across Languages
NAACL 2012 — Montreal Spence Green* Nicholas Andrews# Matthew R. Gormley# Mark Dredze# Christopher D. Manning* *Stanford University
#Johns Hopkins University
Entity Clustering Across Languages NAACL 2012 Montreal Spence - - PowerPoint PPT Presentation
Entity Clustering Across Languages NAACL 2012 Montreal Spence Green* Nicholas Andrews # Matthew R. Gormley # Mark Dredze # Christopher D. Manning* *Stanford University # Johns Hopkins University One Entity, Many Names Qaddafi, Muammar
NAACL 2012 — Montreal Spence Green* Nicholas Andrews# Matthew R. Gormley# Mark Dredze# Christopher D. Manning* *Stanford University
#Johns Hopkins University
Qaddafi, Muammar Al-Gathafi, Muammar al-Qadhafi, Muammar Al Qathafi, Mu’ammar Al Qathafi, Muammar El Gaddafi, Moamar El Kadhafi, Moammar El Kazzafi, Moamer El Qathafi, Mu’Ammar
3 / 41
Qaddafi, Muammar Al-Gathafi, Muammar al-Qadhafi, Muammar Al Qathafi, Mu’ammar Al Qathafi, Muammar El Gaddafi, Moamar El Kadhafi, Moammar El Kazzafi, Moamer El Qathafi, Mu’Ammar
Qaddafi, Muammar Al-Gathafi, Muammar al-Qadhafi, Muammar Al Qathafi, Mu’ammar Al Qathafi, Muammar El Gaddafi, Moamar El Kadhafi, Moammar El Kazzafi, Moamer El Qathafi, Mu’Ammar
Cluster co-referent entity mentions across a corpus (documents and languages)
4 / 41
Cluster co-referent entity mentions across a corpus (documents and languages) Clustering/disambiguation relies on:
◮ Mention similarity ◮ Context similarity
4 / 41
Qaddafi, Muammar Al-Gathafi, Muammar al-Qadhafi, Muammar Al Qathafi, Mu’ammar Al Qathafi, Muammar El Gaddafi, Moamar El Kadhafi, Moammar El Kazzafi, Moamer El Qathafi, Mu’Ammar
6 / 41
6 / 41
The Apple chief executive was former Beatles road manager Neil Aspinall... Sentential context is usually required
7 / 41
Peter said to himself ...
8 / 41
doc1: Peter Jones said ... doc2: I told Mr. Jones ...
[Bagga and Baldwin, 1998] [Baron and Freedman, 2008]
Peter said to himself ...
9 / 41
doc1: Peter Jones said ... doc2: I told Mr. Jones ...
[Bagga and Baldwin, 1998] [Baron and Freedman, 2008] [McNamee et al., 2011]
Peter said to himself ... Peter Jones Peter
[Rao et al., 2011]
10 / 41
doc1: Peter Jones said ... doc2: I told Mr. Jones ...
[Bagga and Baldwin, 1998] [Baron and Freedman, 2008] [McNamee et al., 2011]
Peter said to himself ... Peter Jones Peter
doc1: Peter Jones said ... doc2:
11 / 41
The Apple chief executive was former Beatles road manager Neil Aspinall...
12 / 41
Crisis management Arab Spring (2011)
◮ French, Arabic dialects ◮ Facebook, Twitter, blog...
Haiti earthquake (2010)
◮ Kreyol, English, French ◮ SMS, Twitter, blog...
13 / 41
Crisis management Arab Spring (2011)
◮ French, Arabic dialects ◮ Facebook, Twitter, blog...
Haiti earthquake (2010)
◮ Kreyol, English, French ◮ SMS, Twitter, blog...
Other applications: machine translation, name search
13 / 41
14 / 41
mi = Muammar Qaddafi mj = Moamer El Kazzafi sim(mi, mj) =?
17 / 41
mi = Muammar Qaddafi mj = Moamer El Kazzafi sim(mi, mj) =? Algorithm: Sorted Jaro-Winkler
[Christen, 2006]
17 / 41
mi = map(Apple) = abbl mj = map( ) = abl sim(mi, mj) =?
18 / 41
mi = map(Apple) = abbl mj = map( ) = abl sim(mi, mj) =? Algorithm: Phonetic mapping + classification
18 / 41
Training: parallel name list
◮ 97.1% accuracy on a held-out set
Phonetic mapping: think of Soundex Best features: character bigrams
19 / 41
Method Resource-level Machine Translation High Lexicon Medium Polylingual Topic Model Low
21 / 41
Method Resource-level Machine Translation High Lexicon Medium Polylingual Topic Model Low
◮ MT: Phrasal with NIST-09 data
[Galley et al., 2009]
◮ Lexicon: 31k entries (web and LDC sources)
21 / 41
Polylingual Topic Model (PLTM)
[Mimno et al., 2009]
◮ Words linked through cross-lingual priors ◮ Training: Wikipedia document tuples
Map context words to 1-best topics
22 / 41
[Mimno et al., 2009]
23 / 41
The Apple chief executive was former Beatles road manager Neil Aspinall...
2 1 14 99 7 7 103 79
24 / 41
Bag of words / smoothed unigram distributions Measure: Jensen-Shannon divergence
25 / 41
Two algorithms:
Setup:
◮ Mention similarity as a hard constraint ◮ Cluster distance: context similarity
27 / 41
context context context
28 / 41
context context context
0.40 0.31 0.15
29 / 41
context context context
0.31
30 / 41
Automatic Content Extraction (ACE) 2008 Arabic-English We annotated 216 cross-lingual entities Genres:
conversation
32 / 41
Docs Tokens Entities Chains Mentions Arabic 412 178k 2.6k 4.2k 9.2k English 414 246k 2.3k 4.0k 9.1k
◮ Chain – set of mentions (within-doc) ◮ Entity – set of chains
33 / 41
Our models cluster chains Evaluation:
◮ Gold chains ◮ Predicted chains from SERIF
[Ramshaw et al., 2011]
34 / 41
B3 B3 (cross-lingual only) 50 60 70 80 90 100 85.4 78.8 80.4 66.4 77.3 58.4 70.1 54.5 MT Lexicon PLTM Baseline
35 / 41
B3 B3 (cross-lingual only) 50 60 70 80 90 100 85.4 78.8 80.4 66.4 77.3 58.4 70.1 54.5 MT Lexicon PLTM Baseline
In paper: CEAF, NVI
35 / 41
36 / 41
Correct tony blair
khaled mashaal
Incorrect NSA
CIA
(En and Ar mentions)
republican party
37 / 41
Correct NSA
CIA labour party
Incorrect hamed bin khalifa al- thani
khlfa thny
38 / 41
doc1: Peter Jones said ... doc2: I told Mr. Jones ...
[Bagga and Baldwin, 1998] [Baron and Freedman, 2008] [McNamee et al., 2011]
Peter said to himself ... Peter Jones Peter
doc1: Peter Jones said ... doc2:
39 / 41
Code and corpus: spencegreen.com
Pairwise models don’t scale
◮ See [Rao et al., 2010] and [Singh et al., 2011]
Model at mention level
◮ See Nick Andrews’ talk at EMNLP!
Unified similarity measure
◮ Logistic regression did not generalize
41 / 41