breaking nli systems
play

Breaking NLI Systems with Sentences that Require Simple Lexical - PowerPoint PPT Presentation

Breaking NLI Systems with Sentences that Require Simple Lexical Inferences Max Glockner 1 , Vered Shwartz 2 and Yoav Goldberg 2 1 TU Darmstadt 2 Bar-Ilan University July 18, 2018 SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural


  1. Breaking NLI Systems with Sentences that Require Simple Lexical Inferences Max Glockner 1 , Vered Shwartz 2 and Yoav Goldberg 2 1 TU Darmstadt 2 Bar-Ilan University July 18, 2018

  2. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  3. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  4. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  5. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  6. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  7. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street ⇒ NEUTRAL NEUTRAL Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  8. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street ⇒ NEUTRAL NEUTRAL 3. A magician performing for an audience in a nightclub Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  9. SNLI [Bowman et al., 2015] A large scale dataset for NLI (Natural Language Inference; Recognizing Textual Entailment [Dagan et al., 2013]) Premises are image captions, hypotheses generated by crowdsourcing workers: Premise Street performer is doing his act for kids Hypotheses 1. A person performing for children on the street ⇒ ENTAILMENT ENTAILMENT 2. A juggler entertaining a group of children on the street ⇒ NEUTRAL NEUTRAL 3. A magician performing for an audience in a nightclub ⇒ CONTRADICTION CONTRADICTION Event co-reference assumption Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 2 / 13

  10. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Classifier Extract Features Premise Hypothesis Encoder Encoder Hypothesis Premise 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  11. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  12. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  13. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet SOTA exceeds human performance... 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  14. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet SOTA exceeds human performance... 1 1 Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  15. Neural NLI Models End-to-end, either sentence-encoding or attention-based Label Label Classifier Classifier Extract Features Extract Features Attention Premise Hypothesis Premise Hypothesis Encoder Encoder Encoder Encoder Hypothesis Premise Hypothesis Premise Lexical knowledge: only from pre-trained word embeddings As opposed to using resources like WordNet SOTA exceeds human performance... 1 1 [Gururangan et al., 2018, Poliak et al., 2018]: by learning “easy clues” Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 3 / 13

  16. Do neural NLI models implicitly learn lexical semantic relations? Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 4 / 13

  17. New Test Set We constructed a new test set to answer this question Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  18. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  19. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  20. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ w ′ is in the SNLI vocabulary and in pre-trained embeddings Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  21. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ w ′ is in the SNLI vocabulary and in pre-trained embeddings Crowdsourcing labels (mostly contradictions!) Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

  22. New Test Set We constructed a new test set to answer this question Premise : sentences from the SNLI training set Hypothesis : Replacing a single term w in the premise with a related term w ′ w ′ is in the SNLI vocabulary and in pre-trained embeddings Crowdsourcing labels (mostly contradictions!) Contradiction The man is holding a saxophone → The man is holding an electric guitar Max Glockner, Vered Shwartz and Yoav Goldberg · Breaking NLI Systems with Sentences that Require Simple Lexical Inferences 5 / 13

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend