Bidirectional LSTM-CRF Models for Sequence Tagging
ADVISOR: JIA-LING, KOH SOURCE: CORR 2015 SPEAKER: SHAO-WEI, HUANG DATE: 2020/01/15
for Sequence Tagging ADVISOR: JIA-LING, KOH SOURCE: CORR 2015 - - PowerPoint PPT Presentation
1 Bidirectional LSTM-CRF Models for Sequence Tagging ADVISOR: JIA-LING, KOH SOURCE: CORR 2015 SPEAKER: SHAO-WEI, HUANG DATE: 2020/01/15 2 OUTLINE Introduction Method Experiment Conclusion INTRODUCTION 3 Sequence
ADVISOR: JIA-LING, KOH SOURCE: CORR 2015 SPEAKER: SHAO-WEI, HUANG DATE: 2020/01/15
(B-ORG) (O) (B-MISC) (O) (O) (O) (B-MISC) (O)
Simple RNN model
x(t) h(t-1) h(t) y(t)
Input
LSTM model
𝑢
前一時間點 的hidden
Note:
is the element- wise product.
https://www.itread01.com/content/1545027542.html 𝑢
LSTM model
Bi-LSTM model
CRF model
The score of the 𝑧𝑗𝑢ℎ tag of the 𝑗𝑢ℎ word in a sentence(independently).
0.7 0.1 0.1 0.1 0.1 0.1 0.1 0.7 0.1 0.7 0.1 0.1
W1 W2 W3
tag1 tag2 tag3 tag4
A matrix of transition scores,𝐵𝑧𝑗,𝑧𝑗+1 represents the score of a transition from the tag 𝑧𝑗 to tag 𝑧𝑗+1.
tag1 tag2 tag3 tag4 tag1 tag2 tag3 tag4
0.6 0.2 0.1 0.1 0.1 0.1 0.1 0.7 0.1 0.7 0.1 0.1 0.5 0.1 0.1 0.3
Max
Reference:Attention is all you need
https://blog.csdn.net/jiaowoshouzi/article/details/89 073944
Reference:Attention is all you need
Reference:Transfer learning for scientific data chain extraction in small chemical corpus with BERT-CRF model
EU rejects German call B-ORG O B-MISC O
➢ Penn TreeBank (PTB) :POS tagging ➢ CoNLL 2000: chunking ➢ CoNLL 2003 :named entity tagging
➢ Spelling features ➢ Context features:uni-gram、bi-gram、tri-gram ➢ Word embedding:Senna word enbedding(each word
➢ Comparison with other networks:
Accuracy F1 F1
➢ Performance with only word feature:
Accuracy F1 F1
➢ Systematically compare the performance of aforementioned
➢ The first to apply a bidirectional LSTM CRF model to NLP
➢ Show that BI-LSTMCRF model is robust and it has less
➢ BERT+CRF model(proposed in another paper).