Deep Learning for Language Understanding (at Google Scale) Anjuli - - PowerPoint PPT Presentation

deep learning for language understanding at google scale
SMART_READER_LITE
LIVE PREVIEW

Deep Learning for Language Understanding (at Google Scale) Anjuli - - PowerPoint PPT Presentation

Deep Learning for Language Understanding (at Google Scale) Anjuli Kannan Software Engineer, Google Brain Confidential + Proprietary Confidential + Proprietary Text is just a sequence of words ["hi", "team",


slide-1
SLIDE 1

Confidential + Proprietary Confidential + Proprietary

Deep Learning for Language Understanding (at Google Scale)

Anjuli Kannan Software Engineer, Google Brain

slide-2
SLIDE 2

Confidential + Proprietary

Text is just a sequence of words

["hi", "team", "the", "server", "appears", "to", "be", "dropping", "about", "10%", …]

slide-3
SLIDE 3

About me

  • My team: Google Brain

○ "Make machines intelligent, improve people's lives." ○ Research + software + applications ○ g.co/brain

  • My work is at boundary of research and applications
  • Focus on natural language understanding
slide-4
SLIDE 4

Neural network basics

slide-5
SLIDE 5

Confidential + Proprietary

Neural network

Is a 4 Is a 5 ... ...

Image: Wikipedia

slide-6
SLIDE 6

Confidential + Proprietary

Neural network

Neuron

Is a 4 Is a 5

slide-7
SLIDE 7

Confidential + Proprietary

Basic building block is the neuron

Greg Corrado

slide-8
SLIDE 8

Gradient descent

w’ = w - α ∂wL(w)

w w’

Learning Rate

Slide: Vincent Vanhoucke

slide-9
SLIDE 9

Recurrent neural networks

slide-10
SLIDE 10

Confidential + Proprietary

Recurrent neural networks can model sequences

slide-11
SLIDE 11

Recurrent neural networks can model sequences

How Message

slide-12
SLIDE 12

How are Message

Recurrent neural networks can model sequences

slide-13
SLIDE 13

How are you Message

Recurrent neural networks can model sequences

slide-14
SLIDE 14

How are you ? Message

Recurrent neural networks can model sequences

slide-15
SLIDE 15

Internal state is a fixed length encoding of the message

How are you ? Message

Recurrent neural networks can model sequences

slide-16
SLIDE 16

Sequence-to-sequence models

slide-17
SLIDE 17

Suppose we want to generate email replies

Smartreply Incoming email Response email

slide-18
SLIDE 18

Sequence-to-sequence model

Sutskever et al, NIPS 2014

slide-19
SLIDE 19

Sequence-to-sequence model

encoder decoder

slide-20
SLIDE 20

Sequence-to-sequence model

Ingests incoming message Generates reply message

slide-21
SLIDE 21

Encoder ingests the incoming message

Internal state is a fixed length encoding of the message

How are you ? Message

slide-22
SLIDE 22

Decoder is initialized with final state of encoder

How are you ? __ How are you ? Message

slide-23
SLIDE 23

Decoder is initialized with final state of encoder

How are you ? __ How are you ? Message

slide-24
SLIDE 24

How are you ? __ I Message Response

Decoder predicts next word

slide-25
SLIDE 25

How are you ? __ I I am Message Response

Decoder predicts next word

slide-26
SLIDE 26

How are you ? __ I am I am great Message Response

Decoder predicts next word

slide-27
SLIDE 27

How are you ? __ I am great I am great ! Message Response

Vinyals & Le, ICML DL 2015 Kannan et al, KDD 2016

Decoder predicts next word

slide-28
SLIDE 28

What the model can do

slide-29
SLIDE 29

What the model can do

slide-30
SLIDE 30

Summary

  • Neural networks learn feature representations from raw data
  • Recurrent neural networks have statefulness which allows them to model

sequences of data such as text

  • The sequence-to-sequence model contains two recurrent neural networks: one

to encode an input sequence and one to generate an output sequence

slide-31
SLIDE 31

Smartreply

slide-32
SLIDE 32

Google Translate

slide-33
SLIDE 33

Research: Speech recognition

slide-34
SLIDE 34

Research: Electronic health records

slide-35
SLIDE 35

What's next?

?

slide-36
SLIDE 36

Resources

  • All tensorflow tutorials:

https://www.tensorflow.org/versions/master/tutorials/index.html

  • Sequence-to-sequence tutorial (machine translation):

https://www.tensorflow.org/versions/master/tutorials/seq2seq

  • Chris Olah's blog: http://colah.github.io/
slide-37
SLIDE 37

Thank you!