Neural Conversational Models Human: What is the purpose of living? - - PowerPoint PPT Presentation

β–Ά
neural conversational models
SMART_READER_LITE
LIVE PREVIEW

Neural Conversational Models Human: What is the purpose of living? - - PowerPoint PPT Presentation

Neural Conversational Models Human: What is the purpose of living? Machine: To live forever. Berkay Antmen March 8, 2016 Conversational model Purpose: Given previous sentences of the dialogue and context, output a response Why?


slide-1
SLIDE 1

Neural Conversational Models

Berkay Antmen March 8, 2016

Human: What is the purpose of living? Machine: To live forever.

slide-2
SLIDE 2

Conversational model

  • Purpose: Given previous sentences of the dialogue and context,
  • utput a response
slide-3
SLIDE 3
  • Why?
  • goal driven dialogue systems
  • e.g. tech support
  • non-goal driven dialogue systems
  • e.g. language learning, video game characters
  • How?
  • discriminative
  • generative
  • heavily hand-crafted
  • data-driven systems
slide-4
SLIDE 4

Demo (Cleverbot)

  • http://www.cleverbot.com/
  • http://www.cleverbot.com/conv/201603150055/VWU01366204_Hi-can-you-

help-me (Troubleshooting)

  • http://www.cleverbot.com/conv/201603150111/VWU01366307_Hello (Basic)
  • http://www.cleverbot.com/conv/201603150120/VWU01366357_What-is-the-

purpose-of-life (Philosophical)

  • http://www.cleverbot.com/conv/201603150204/VWU01366635_We-are-no-

strangers-to-love (extra)

slide-5
SLIDE 5

Frameworks

  • sequence-to-sequence (seq2seq)
  • classification problem over a known vocabulary
  • input: sequence of tokens
  • output: sequence of tokens

image: Sutskever et. al. 2015

slide-6
SLIDE 6

Frameworks: seq2seq

  • The goal: estimate
  • problem: boundaries
  • solution:
  • training: maximize (target given source)
  • inference:
  • approximated by beam search

equation images: Sutskever et. al. 2015

slide-7
SLIDE 7

Beam Search

image: http://bit.ly/251bIfl

w=3

slide-8
SLIDE 8

A Neural Conversational Model

  • IT helpdesk dataset of conversations (closed-domain)
  • OpenSubtitles movie transcript dataset (open-domain)
  • Experiments: troubleshooting, general knowledge, philosophical etc.
slide-9
SLIDE 9
  • training: maximize cross entropy of the correct sequence given its context
  • (aside) how is cross entropy measured when the true distribution of the words in the

corpus is not known? Monte Carlo estimation: training set is treated as samples from the true distribution

  • inference: greedy search

A Neural Conversational Model

image: Chris Olah

slide-10
SLIDE 10

Some results (troubleshooting)

Password issues Browser issues http://www.cleverbot.com/conv/201603150055/VWU0136620 4_Hi-can-you-help-me Cleverbot:

slide-11
SLIDE 11

Some more results

Basic Contexts and multiple choice

Cleverbot: http://www.cleverbot.com/conv/201603150111/VWU01366307_Hello

slide-12
SLIDE 12

Some more results

Philosophical Opinions

http://www.cleverbot.com/conv/201603150120/VWU0136635 7_What-is-the-purpose-of-life Cleverbot:

slide-13
SLIDE 13

Evaluation

  • Perplexity measures how well a model predicts the given samples
  • 2πΌπ‘Ÿ(𝑇1,β€¦π‘‡π‘œ) = 2βˆ’ 𝑗 π‘Ÿ 𝑇𝑗 log2 π‘Ÿ(𝑇𝑗)

Experiment Model Perplexity IT Helpdesk Troubleshooting N-grams 18 IT Helpdesk Troubleshooting Neural conversational model 8 OpenSubtitles N-grams 28 OpenSubtitles Neural conversational model 17

slide-14
SLIDE 14

Evaluation

  • human evaluation against a rule-based bot (CleverBot)
  • asked a list of questions to both models
  • judges picked the bot they preferred
  • Mechanical Turk

# questions # judges # prefer neural model # prefer CleverBot # tie # disagreement 200 4 97 60 20 23

slide-15
SLIDE 15

Wrong objective function?

  • the answers are not diverse, i.e. likely to give most probable answers

without giving out much information

  • e.g. S=β€œHow old are you?” T=β€œI don’t know.”
  • π‘ž(π‘ˆ|𝑇) high, π‘ž(𝑇|π‘ˆ) low
  • e.g. S=β€œHow old are you?” T=β€œI am 10 years old”
  • π‘ž(π‘ˆ|𝑇) lower, π‘ž(𝑇|π‘ˆ) higher
  • not really obvious from the selected examples in the paper
slide-16
SLIDE 16

A Diversity-Promoting Objective Function for Neural Conversation Models

Li et. al. 2015

slide-17
SLIDE 17

A Diversity-Promoting Objective Function for Neural Conversation Models

  • An alternative objective function: Maximum Mutual Information (MMI)
  • maximize mutual information between source (S) and target (T)
  • 𝐽 𝑇, π‘ˆ = log( π‘ž(𝑇,π‘ˆ)

π‘ž 𝑇 π‘ž(π‘ˆ))

  • π‘ˆ = argπ‘ˆ max π‘šπ‘π‘•π‘ž π‘ˆ 𝑇 βˆ’ πœ‡π‘šπ‘π‘•π‘ž(π‘ˆ)
  • remember, previously
slide-18
SLIDE 18

Some results (OpenSubtitles)

slide-19
SLIDE 19

Some results (Twitter)

slide-20
SLIDE 20
  • Hierarchical Recurrent Encoder Decoder (HRED)

Frameworks

image: Serban et. al. 2015

slide-21
SLIDE 21

Frameworks: HRED

  • Motivation?
slide-22
SLIDE 22

Hierarchical Neural Network Generative Models for Movie Dialogues

  • Non-goal driven: can be easily adapted to specific tasks
  • Bootstrapping
  • from word embeddings OR
  • from a large non-dialogue corpus (Q-A SubTle containing 5.5 pairs)
  • Interactive dialogue structure
  • end-of-utterance token
  • continued-utterance token
slide-23
SLIDE 23

Dataset

  • why movie scripts?
  • large dataset
  • wide range of topics
  • long dialogues with few participants
  • relatively few spelling mistakes and acronyms
  • similar to human spoken conversations
  • mostly single dialogue thread
  • atomic entries are triples
  • 13M words total; 10M in training
slide-24
SLIDE 24

Evaluations (movie dialogue generation)

  • test set perplexity and classification errors when bootstrapping from

SubTle corpus

slide-25
SLIDE 25

Evaluations

slide-26
SLIDE 26

Future work?

  • study larger length dialogues (as opposed to triplets)
  • bootstrapping on other non-dialogue but large datasets
slide-27
SLIDE 27

Thank you!

Questions?

slide-28
SLIDE 28

References

  • seq2seq http://arxiv.org/abs/1409.3215
  • neural conversational http://arxiv.org/abs/1506.05869
  • hierarchical http://arxiv.org/abs/1507.02221
  • hierarchical conversational http://arxiv.org/abs/1507.04808
  • MMI http://arxiv.org/abs/1510.03055