VoxEL: A Benchmark Dataset for Multilingual Entity Linking Henry - - PowerPoint PPT Presentation

voxel a benchmark dataset for
SMART_READER_LITE
LIVE PREVIEW

VoxEL: A Benchmark Dataset for Multilingual Entity Linking Henry - - PowerPoint PPT Presentation

VoxEL: A Benchmark Dataset for Multilingual Entity Linking Henry Rosales-M endez, Aidan Hogan and Barbara Poblete University of Chile { hrosales,ahogan,bpoblete } @dcc.uchile.cl October 10, 2018 ISWC 2018 - The 17th Internationl


slide-1
SLIDE 1

VoxEL: A Benchmark Dataset for Multilingual Entity Linking †

Henry Rosales-M´ endez, Aidan Hogan and Barbara Poblete

University of Chile {hrosales,ahogan,bpoblete}@dcc.uchile.cl

October 10, 2018

† ISWC 2018 - The 17th Internationl Semantic Web Conference

slide-2
SLIDE 2

Example

slide-3
SLIDE 3

Example - Entity Recognition

slide-4
SLIDE 4

Example - Entity Disambiguation

slide-5
SLIDE 5

Applications

  • Semantic Search
  • Semantic Annotations
  • Relation Extraction
  • Topic Extraction
slide-6
SLIDE 6

Name Variations in Entity Linking

Michael J. Jackson King of Pop Michael Joseph Jackson

slide-7
SLIDE 7

Name Variations in Entity Linking

Michael Jackson

slide-8
SLIDE 8

Multilingual Entity Linking - English

slide-9
SLIDE 9

Multilingual Entity Linking - Italian

slide-10
SLIDE 10

Multilingual Entity Linking - Spanish

slide-11
SLIDE 11

Datasets

slide-12
SLIDE 12

Datasets

slide-13
SLIDE 13

Goals

1 Create a benchmark dataset for multilingual Entity Linking

slide-14
SLIDE 14

Curated source: VoxEurop

slide-15
SLIDE 15

Example - Any other entity?

slide-16
SLIDE 16

Example - Any other entity?

slide-17
SLIDE 17

Example - Any other entity?

slide-18
SLIDE 18

Example - Any other entity?

slide-19
SLIDE 19

Example annotations produced by four EL systems

slide-20
SLIDE 20

Example annotations produced by four EL systems

slide-21
SLIDE 21

Example annotations produced by four EL systems

Aida

slide-22
SLIDE 22

Example annotations produced by four EL systems

Aida Babelfy

slide-23
SLIDE 23

Example annotations produced by four EL systems

Aida Babelfy

DBpedia Spotlight

slide-24
SLIDE 24

Example annotations produced by four EL systems

Aida Babelfy

DBpedia Spotlight T agME

slide-25
SLIDE 25
  • What should Entity Linking link?
slide-26
SLIDE 26

Datasets

slide-27
SLIDE 27

Datasets

slide-28
SLIDE 28

Goals

1 Create a benchmark dataset for multilingual Entity Linking

slide-29
SLIDE 29

Goals

1 Create a benchmark dataset for multilingual Entity Linking 2 Create two versions of the dataset: strict and relaxed.

slide-30
SLIDE 30

Strict verison: class-based definition

slide-31
SLIDE 31

Strict verison: class-based definition

slide-32
SLIDE 32

Strict verison: class-based definition

?

slide-33
SLIDE 33

Relaxed version: Knowledge Base definition

slide-34
SLIDE 34

Creation of VoxEL dataset

  • It is based on curated text from five languages.
  • Same sentences by each corresponding document.
  • Same annotations by each corresponding sentence.
  • Revision process.
slide-35
SLIDE 35

Summary

slide-36
SLIDE 36

Summary

slide-37
SLIDE 37

Summary

slide-38
SLIDE 38

Experiments

1 GERBIL Evaluation of state-of-the-art approaches

slide-39
SLIDE 39

Experiments

slide-40
SLIDE 40

Experiments

slide-41
SLIDE 41

Experiments

(a) Results of the Relaxed version of VoxEL DB-sp FREME TAGME

.58 .60 .66 .61

Babefy

r s

Babefy

.58

  • Avg. of Micro F1

DE EN ES FR IT

.34 .30 .34.33.32 .65 .60 .50.47 .57 .40 .27 .28 .22 .34 .59 .34 .27 .39 .19

THD (b) Results of the Strict version of VoxEL DB-sp FREME TAGME

.76 .78 .81 .74

Babefy

r s

Babefy

.71

  • Avg. of Micro F1

.64 .70 .72.70.71 .71 .64 .49.51 .65 .81 .54 .60 .53 .74 .86 .75 .60 .72 .50

THD

slide-42
SLIDE 42

Experiments

(a) Results of the Relaxed version of VoxEL DB-sp FREME TAGME

.58 .60 .66 .61

Babefy

r s

Babefy

.58

  • Avg. of Micro F1

DE EN ES FR IT

.34 .30 .34.33.32 .65 .60 .50.47 .57 .40 .27 .28 .22 .34 .59 .34 .27 .39 .19

THD (b) Results of the Strict version of VoxEL DB-sp FREME TAGME

.76 .78 .81 .74

Babefy

r s

Babefy

.71

  • Avg. of Micro F1

.64 .70 .72.70.71 .71 .64 .49.51 .65 .81 .54 .60 .53 .74 .86 .75 .60 .72 .50

THD

slide-43
SLIDE 43

Experiments

(a) Results of the Relaxed version of VoxEL DB-sp FREME TAGME

.58 .60 .66 .61

Babefy

r s

Babefy

.58

  • Avg. of Micro F1

DE EN ES FR IT

.34 .30 .34.33.32 .65 .60 .50.47 .57 .40 .27 .28 .22 .34 .59 .34 .27 .39 .19

THD (b) Results of the Strict version of VoxEL DB-sp FREME TAGME

.76 .78 .81 .74

Babefy

r s

Babefy

.71

  • Avg. of Micro F1

.64 .70 .72.70.71 .71 .64 .49.51 .65 .81 .54 .60 .53 .74 .86 .75 .60 .72 .50

THD

slide-44
SLIDE 44

Experiments

(a) Results of the Relaxed version of VoxEL DB-sp FREME TAGME

.58 .60 .66 .61

Babefy

r s

Babefy

.58

  • Avg. of Micro F1

DE EN ES FR IT

.34 .30 .34.33.32 .65 .60 .50.47 .57 .40 .27 .28 .22 .34 .59 .34 .27 .39 .19

THD (b) Results of the Strict version of VoxEL DB-sp FREME TAGME

.76 .78 .81 .74

Babefy

r s

Babefy

.71

  • Avg. of Micro F1

.64 .70 .72.70.71 .71 .64 .49.51 .65 .81 .54 .60 .53 .74 .86 .75 .60 .72 .50

THD

slide-45
SLIDE 45

Experiments

1 GERBIL Evaluation of state-of-the-art approaches

slide-46
SLIDE 46

Experiments

1 GERBIL Evaluation of state-of-the-art approaches 2 Evaluate the performance of state-of-the-art approaches using

machine translation.

slide-47
SLIDE 47

Experiments

DE EN ES FR IT DE EN ES FR IT Input T ext System Configuration

slide-48
SLIDE 48

Experiments

DE EN ES FR IT DE EN ES FR IT Input T ext System Configuration

slide-49
SLIDE 49

Experiments

DE EN ES FR IT DE EN ES FR IT Input T ext System Configuration

slide-50
SLIDE 50

Experiments

DE EN ES FR IT DE EN ES FR IT Input T ext System Configuration

slide-51
SLIDE 51

Experiments

(a) Results of the Relaxed version of VoxEL DB-sp FREME TAGME

.51 .52 .55

Babefy

r s

Babefy

  • Avg. of Micro F1

Calibrated Translation English

.31 .30 .31 .45 .33 .47 .40 .31 .41 .43 .41 .46 .39 .24

THD

.39

(b) Results of the Strict version of VoxEL DB-sp FREME TAGME

.30 .35 .32

Babefy

r s

Babefy

  • Avg. of Micro F1

.53 .60 .57 .71 .53 .71 .70 .57 .71 .33 .27 .33 .60 .55

THD

.63

slide-52
SLIDE 52

Conclusion

Our main contribution is VoxEL (https://dx.doi.org/10.6084/m9.figshare.6539675)

  • Most systems perform (much) better for English.
  • Machine Translation could be an option to address

multilingual domains in Entity Linking.

slide-53
SLIDE 53

Poster P20: Machine Translation vs. Multilingual Approaches for Entity Linking

EN IT ES

slide-54
SLIDE 54

VoxEL: A Benchmark Dataset for Multilingual Entity Linking †

Henry Rosales-M´ endez, Aidan Hogan and Barbara Poblete

University of Chile {hrosales,ahogan,bpoblete}@dcc.uchile.cl

October 10, 2018

† ISWC 2018 - The 17th Internationl Semantic Web Conference