multi level representations for fine grained typing of
play

Multi-level Representations for Fine-Grained Typing of Knowledge Base - PowerPoint PPT Presentation

Multi-level Representations for Fine-Grained Typing of Knowledge Base Entities Yadollah Yaghoobzadeh and Hinrich Schtze LMU Munich, Germany Presented by: Xiaotao Gu Knowledge Graph/Base Image retrieved from:


  1. Multi-level Representations for Fine-Grained Typing of Knowledge Base Entities Yadollah Yaghoobzadeh and Hinrich Schütze LMU Munich, Germany Presented by: Xiaotao Gu

  2. Knowledge Graph/Base Image retrieved from: https://www.ambiverse.com/knowledge-graphs-encyclopaedias-for-machines/

  3. Entity Typing candidate types City entity typing State Supervised Barack Obama Classification! Person Country How to extract high quality feature representation? classifier feature label

  4. Joint Representation

  5. Character-level Representation Character-level Representation Max Pooling concatenate Convolution Filter of size 2 Layer (3 * 6 feature map) Filter of size 4 (4 * 4 feature map) Character Embedding Character Embedding of “S” + Lookup Table s p a n i s h

  6. Joint Representation

  7. Word-level Representation Word embedding based on Distributional Hypothesis Semantic meaning of words are useful for typing! E.g. “XXX Lake ” implies $LOCATION Entity embedding = average word embedding E( Lake Michigan ) = 0.5 * { E( Lake ) + E( Michigan ) } Subword/morphology in words are useful for typing! E.g. “Span ish ” implies $LANGUAGE FastText : n-gram subword embedding! FastText: https://research.fb.com/fasttext/ Image retrieved from: Professor Julia Hockenmaier’s slides for CS447: Natural Language Processing

  8. Joint Representation

  9. Entity-level Representation Corpus+SkipGram Unique Identifier Entity ( Lake Michigan ) Identifier ( Lake_Michigan ) Entity Embedding Capture Context Information

  10. Entity-level Representation

  11. Joint Representation

  12. Experiment Task: Entity Typing Dataset: FIGMENT ● 102 types ● 200K Freebase entities, 60K for testing ● 12K head entities (freq > 100), 10K tail entities (freq < 5) Evaluation Metrics ● Accuracy : correct iff all types of an entity are inferred correct and no wrong types are inferred ● Micro average F1: F1 of all type-entity assignment decisions ● Macro average F1: F1 of types assigned to an entity, averaged over entities

  13. Experiment baseline Character-level

  14. Experiment baseline Character-level Word-level

  15. Experiment baseline Character-level Word-level Word + Character

  16. Experiment baseline Character-level Word-level Word + Character Entity-level

  17. Experiment baseline Character-level Word-level Word + Character Entity-level Entity + Word + Character

  18. Experiment Character-Level: CNN performs best for capturing local features Word-Level: Subword improves the performance, especially for rare entities Joint entity-word-character information: achieves the best performance External Information: adding description text helps tail entities! Add description

  19. Summary High-quality Entity Representation from: ● Context information: from large corpus ● Surface name information: word, subword, character-level information ● External knowledge: description, relational links ● The joint representation is the most informative for entity typing. Thank you!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend