SLIDE 1 REET Joint Relation Extraction and Entity Typing via Multi-task Learning
ADVISOR: JIA-LING, KOH SOURCE: NLPCC 2019 SPEAKER: SHAO-WEI, HUANG DATE: 2020/03/16
1
SLIDE 2
⚫ Introduction ⚫ Method ⚫ Experiment ⚫ Conclusion
2
OUTLINE
SLIDE 3
INTRODUCTION
➢ Relation Extraction(RE):Extracting semantic
relations between two entities from the text corpus.
3 (Ex): Steve jobs was the co-founder of apple. Co-Founder
SLIDE 4
INTRODUCTION
➢ Entity Typing(ET):Assign types into the entity
mention in a sentence.
4 (Ex): Steve jobs was the co-founder of apple. Person Company
SLIDE 5
INTRODUCTION
➢ Most existing works solve RE and ET separately
and regard them as independent tasks.
5
➢ In fact, the two tasks have a strong inner
relationship.
REET Model Joint Relation Extraction and Entity Typing.
SLIDE 6 6
INTRODUCTION
Problem definition
➢ Given a sentence s = {𝑥1, 𝑥2, … 𝑓1, … , 𝑓2 …} and
two target entities (𝑓1, 𝑓2).
➢ Subtasks:
- 1. Relation extraction for the entity pair.
- 2. Entity typing for 𝑓1.
- 3. Entity typing for 𝑓2.
SLIDE 7
OUTLINE
Introduction Method Experiment Conclusion
7
SLIDE 8
8
FRAMEWORK
8
Entity Typing Relation Extraction
SLIDE 9 METHOD
9
Relation Extraction Module
➢For a sentence s = {𝑥1, 𝑥2, … 𝑓1, … , 𝑓2 … 𝑥𝑜} , transform each word 𝑥𝑗 into: (Ex): Steve jobs was the co-founder of apple.
𝑥𝑗
2
- 3
- 1. Word embeddings
- 2. Position embeddings :Encodes the relative
distances between 𝑥𝑗 and the two entities.
SLIDE 10 METHOD
10
Relation Extraction Module
➢Convolution and Piecewise max pooling :
Steve Jobs was the co-founder
apple
Word + Position Embeddings
0.7 0.8 0.9 0.2 0.5 0.6
Filters
0.7 0.8 0.9 0.2 0.5 0.6
Input Convolution
0.7 0.9 0.6
Piecewise max pooling f1 f2 f3 f4 Tanh
Sentence representation S
SLIDE 11 11
METHOD Entity Typing Module
➢Input Layer:Shared with RE module.
- 1. Word embeddings.
- 2. Position embeddings.
11 ➢Bi-LSTM layer:Obtain the hidden state(high-level semantic representation) of each 𝑥𝑗.
SLIDE 12 12
METHOD Entity Typing Module
➢Couple Attention:To get entity-related representations for sentences. 12
⚫ ⚫
Treat entities as query, other words as key.
entity1, entity2 The weights of the i-th word under the m-th entity
⚫
Weight sum:
⚫
Final representation
SLIDE 13 13
METHOD Multi-task Learning Framework
➢ REET1:Treat RE task and ET task are independent and only share input embedding layers.
⚫
Prediction probabilities for RE and ET respectively.
⚫
S 𝐔𝟐 𝐔𝟑
SLIDE 14 14
METHOD Multi-task Learning Framework
➢ REET2 : Concatenate representations
- f RE and ET before the last classication
layer.
⚫
S 𝐔𝟐 𝐔𝟑
⚫
Prediction probabilities for RE and ET respectively. ***RE and ET can share a high-level feature with each other.
SLIDE 15 15
METHOD Multi-task Learning Framework
➢ Loss Function:Cross entropy loss.
⚫
➢ Multi-task Learning:Add the loss of each task together.
⚫
Balance weight
SLIDE 16
OUTLINE
Introduction Method Experiment Conclusion
16
SLIDE 17 17
EXPERIMENT
Dataset
➢ NYT+Freebase:Aligning entities and relations in Freebase with the corpus
➢ Google Distant Supervision(SGD): Extracted from Google Relation
Extraction corpus and is a human-judged dataset.
https://ai.googleblog.com/2013/04/50000-lessons-
- n-how-to-read-relation.html
SLIDE 18 18
EXPERIMENT
Performance in RE
https://blog.csdn.net/u013249853/article/details/961 32766
SLIDE 19
19
EXPERIMENT
Performance in ET
SLIDE 20 20
EXPERIMENT
Parameter analysis
https://blog.xuite.net/metafun/life/65137005- Information+Retrieval%E7%9A%84%E8%A1%A1%E9%8 7%8F%E6%8C%87%E6%A8%99-MAP%E5%92%8CMRR
***
SLIDE 21
OUTLINE
Introduction Method Experiment Conclusion
21
SLIDE 22 22
CONCLUSION
➢ Propose a multi-task learning frame that integrates relation
extraction task and entity typing task jointly.
➢ The two tasks share low-level (i.e., input embedding layer)
and high-level information (i.e., task-specic feature).
➢ Both relation extraction task and entity typing task achieve a
signicant improvement and our approach outperforms many baseline methods.