Discourse Marker Augmented Network with Reinforcement Learning for - - PowerPoint PPT Presentation

discourse marker augmented network with reinforcement
SMART_READER_LITE
LIVE PREVIEW

Discourse Marker Augmented Network with Reinforcement Learning for - - PowerPoint PPT Presentation

Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference Authors Boyuan Pan, Yazheng Yang, Zhou Zhao, Yueting Zhuang, Deng Cai, Xiaofei He Organization Zhejiang University, China What is Natural


slide-1
SLIDE 1

Discourse Marker Augmented Network with Reinforcement Learning for Natural Language Inference

Authors

Boyuan Pan, Yazheng Yang, Zhou Zhao, Yueting Zhuang, Deng Cai, Xiaofei He

Organization

Zhejiang University, China

slide-2
SLIDE 2

What is Natural Language Inference (NLI)?

Premise Hypothesis

Entailment

slide-3
SLIDE 3

What is Natural Language Inference (NLI)?

?

Premise Hypothesis

Neutral

slide-4
SLIDE 4

What is Natural Language Inference (NLI)?

Premise Hypothesis

Contradiction

slide-5
SLIDE 5

Applications

  • Question Answering
  • Machine Translation
  • Semantic Search
  • Text Summarization

slide-6
SLIDE 6

Discourse Marker

  • A discourse marker is a word or

a phrase that plays a role in managing the flow and structure of discourse.

  • Examples: so, because, and, but, or…
slide-7
SLIDE 7

Discourse Marker & NLI?

But Because If Although And So

Entailment Neutral Contradiction

slide-8
SLIDE 8

Related Works

  • Datasets

SNLI (Bowman et al., 2015) MultiNLI (Williams et al., 2017)

  • SOTA Neural Network Models

CAFE (Tay et al., 2017) KIM (Chen et al., 2017) DIIN (Gong et al., 2018)

slide-9
SLIDE 9

Related Works

  • Transfer Learning for NLI

Skip-thoughts (Vendrov et al., 2016) Cove (McCann et al., 2017)

  • Discourse Marker Applications

DisSent (Nie et al., 2017)

slide-10
SLIDE 10

Discourse Marker Prediction (DMP)

It’s rainy outside but we will not take the umbrella It’s rainy outside + But + We will not take the umbrella (S1, S2) Neural Networks M So Because But … … If

slide-11
SLIDE 11

Discourse Marker Prediction (DMP)

Glove Glove

BiLSTM Sentence Representations Sentence1 Sentence2 Prediction Last hidden state Max pooling over all the hidden states To Be Transferred

slide-12
SLIDE 12

Discourse Marker Augmented Network (NLI Model)

Glove Glove Char Char POS POS NER NER EM EM

BiLSTM Hypothesis

Encoding Layer

Premise

slide-13
SLIDE 13

Discourse Marker Augmented Network (NLI Model)

Glove Glove Char Char POS POS NER NER EM EM

BiLSTM Hypothesis Premise BiLSTM Sentence Representations Pre-trained DMP Model:

slide-14
SLIDE 14

Discourse Marker Augmented Network (NLI Model)

Interaction ------ Similarity Matrix

The i-th word of the premise The j-th word of the hypothesis The sentence representation of the premise The sentence representation of the hypothesis

slide-15
SLIDE 15

Discourse Marker Augmented Network (NLI Model)

Attention Mechanism Prediction Similarity Matrix

Modeling vector of the premise Modeling vector of the hypothesis The sentence representation of the premise The sentence representation of the hypothesis

slide-16
SLIDE 16

Training

Cross Entropy Loss

Correct Label: neutral Original Labels: neutral, neutral, entailment, entailment, neutral

slide-17
SLIDE 17

Training

Previous action policy that predicts the label given P and H.

slide-18
SLIDE 18

Experiments (Datasets)

  • Stanford Natural Language Inference (SNLI) (Bowman et al., 2015)

570k human annotated sentence pairs

  • Multi-Genre Natural Language Inference

(MultiNLI) (Williams et al., 2017) 433k human annotated sentences pairs

  • BookCorpus (Zhu et al., 2015)

6.5M pairs of sentences for 8 discourse markers

slide-19
SLIDE 19

Experiments (Results)

Sentence Encoding- Based Models Other Neural Network Models Ensemble Models

slide-20
SLIDE 20

Experiments (Analysis)

slide-21
SLIDE 21

Experiments (Analysis)

Premise: “3 young man in hoods standing in the middle of a quiet street facing the camera.” Hypothesis: “Three people sit by a busy street bare-headed.

slide-22
SLIDE 22

Conclusion

  • We solve the task of the natural language inference via

transferring knowledge from another supervised task.

  • We propose a new objective function to make full use of

the labels’ information.

  • In the future work, we would like to explore some other

transfer learning sources.

slide-23
SLIDE 23

Thank You !