An Attention-based Model for Joint Extraction of Entities and Relations with Implicit Entity Features
ADVISOR: JIA-LING, KOH SOURCE: WWW 2019 SPEAKER: SHAO-WEI, HUANG DATE: 2019/09/20
An Attention-based Model for Joint Extraction of Entities and - - PowerPoint PPT Presentation
1 An Attention-based Model for Joint Extraction of Entities and Relations with Implicit Entity Features ADVISOR: JIA-LING, KOH SOURCE: WWW 2019 SPEAKER: SHAO-WEI, HUANG DATE: 2019/09/20 2 OUTLINE Introduction Method Experiment
ADVISOR: JIA-LING, KOH SOURCE: WWW 2019 SPEAKER: SHAO-WEI, HUANG DATE: 2019/09/20
Eentiy 1 Eentiy 2 Semantic relation
➢ Pipelined models: ⚫ Identify the entity pair first. And then predict the
Result have an impact on it
➢ Joint models(this paper use):
⚫
Other B:Begin
(Have Begin, Inside, End, Single four type)
PR:President-Of
(Predfined 24 relation type)
1:First entity E:End PR:President-Of 2:Second entity
Features(Word embedding & character embedding )
⚫ Pre-train the embedding for words.
⚫ Each word is broken up into
⚫ Each characters are mapping to
⚫ Adopt Bi-LSTM to generates the
d
a l d Bi- LSTM Bi- LSTM Bi- LSTM Bi- LSTM Bi- LSTM Bi- LSTM ℎ1 ℎ2 ℎ3 ℎ4 ℎ5 ℎ6 Character embedding for the word
Features(Implicit features)
⚫ Pre-train an model on an existing
⚫ Feed the input sentence(Danald
⚫ The hidden vectors are entity features.
https://www.itread01.com/content/1545027542.html 𝑢
Encoder layer
⚫
⚫ Use Bi-LSTM to computes
The input of attention layer
⚫
Attention vector (relevant representation)
⚫ Allow the model to select the
⚫ When predicting the tag of a word,
⚫
Output of attention layer Weight
The input of decoding layer
⚫ Adopt LSTM to generate vectors
t-1-th hidden state
t-1-th tag embedding t-th
Attention layer
⚫ Adopt a softmax classifier to compute
⚫ Objective function:
➢ NYT:
➢ Dimension sizes:
➢ Comparison with baselines:
when all correct.
➢ Ablation results:
when all correct.
➢ Ablation results on triplet elements:
➢ Visualization of attention weights:
➢ Propose an attention-based model enhanced with
➢ This model can take advantage of the entity features
➢ Design a Tag-Aware attention mechanism which enables