recurrent neural networks lstms attention
play

Recurrent Neural Networks + LSTMs + Attention Surag Nair (based on - PowerPoint PPT Presentation

Recurrent Neural Networks + LSTMs + Attention Surag Nair (based on slides by Xavier Gir-i-Nieto, Santi Pascual and M. Malinowski) Multilayer Perceptron The output depends ONLY on the current input. Alex Graves, Supervised Sequence


  1. Recurrent Neural Networks + LSTMs + Attention Surag Nair (based on slides by Xavier Giró-i-Nieto, Santi Pascual and M. Malinowski)

  2. Multilayer Perceptron The output depends ONLY on the current input. Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” 2

  3. Recurrent Neural Network (RNN) The hidden layers and the output depend from previous states of the hidden layers Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” 3

  4. Recurrent Neural Networks (RNN) Each node represents a layer of neurons at a single timestep. t t-1 t+1 Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” 4

  5. Recurrent Neural Networks (RNN) The input is a SEQUENCE x(t) of any length. t t-1 t+1 Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” 5

  6. Recurrent Neural Networks (RNN) Must learn temporally shared weights w2; in addition to w1 & w3. t t-1 t+1 Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” 6

  7. Bidirectional RNN (BRNN) Must learn weights w2, w3, w4 & w5; in addition to w1 & w6. Alex Graves, “Supervised Sequence Labelling with Recurrent Neural Networks” 7

  8. Formulation: Single recurrence One-time Recurrence Slide: Santi Pascual 17

  9. Formulation: Multiple recurrences One time-step recurrence Recurrence T time steps recurrences Slide: Santi Pascual 9

  10. RNN problems Long term memory vanishes because of the T nested multiplications by U. ... Slide: Santi Pascual 10

  11. RNN problems During training, gradients may explode or vanish because of temporal depth. Example: Back- propagation in time with 3 steps. Slide: Santi Pascual 11

  12. Vanishing/Exploding Gradient Problem Pascanu, Razvan, Tomas Mikolov, and Yoshua Bengio. "On the difficulty of training recurrent neural networks." ICML (3) 28 (2013): 1310-1318.

  13. Long Short-Term Memory (LSTM) Hochreiter, Sepp, and Jürgen Schmidhuber. "Long short-term memory." Neural computation 9, no. 8 (1997): 1735-1780. 13

  14. Long Short-Term Memory (LSTM) Based on a standard RNN whose neuron activates with tanh ... Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) 14

  15. Long Short-Term Memory (LSTM) C t is the cell state, which flows through the entire chain... Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) 15

  16. Long Short-Term Memory (LSTM) ...and is updated with a sum instead of a product. This avoid memory vanishing and exploding/vanishing backprop gradients. Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) 16

  17. Long Short-Term Memory (LSTM) Forget Gate : Concatenate Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) / Slide: Alberto Montes 17

  18. Long Short-Term Memory (LSTM) Input Gate Layer New contribution to cell state Classic neuron Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) / Slide: Alberto Montes 18

  19. Long Short-Term Memory (LSTM) Update Cell State (memory): Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) / Slide: Alberto Montes 19

  20. Long Short-Term Memory (LSTM) Output Gate Layer Output to next layer Figure: Cristopher Olah, “Understanding LSTM Networks” (2015) / Slide: Alberto Montes 20

  21. Gated Recurrent Unit (GRU) Similar performance as LSTM with less computation. Cho, Kyunghyun, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. "Learning phrase representations using RNN encoder-decoder for statistical machine translation." arXiv preprint arXiv:1406.1078 (2014). 21

  22. Attention - Motivation • Long term memories - attending to memories ‣ Dealing with gradient vanishing problem • Exceeding limitations of a global representation ‣ Attending/focusing to smaller parts of data - patches in images - words or phrases in sentences • Decoupling representation from a problem ‣ Different problems required different sizes of representations - LSTM with longer sentences requires larger vectors • Overcoming computational limits for visual data ‣ Focusing only on the parts of images ‣ Scalability independent of the size of images • Adds some interpretability to the models (error inspection)

  23. Extension of LSTM via context vector

  24. Soft Attention Example : http://distill.pub/2016/augmented-rnns/

  25. Teaching Machines to Read and Comprehend Karl Moritz Hermann, Tomas Kocisky, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, Lei Yu, and Phil Blunsom pblunsom@google.com

  26. Features and NLP Twenty years ago log-linear models allowed greater freedom to model correlations than simple multinomial parametrisations, but imposed the need for feature engineering.

  27. Features and NLP Distributed/neural models allow us to learn shallow features for our classifiers, capturing simple correlations between inputs.

  28. Deep Learning and NLP Fully connected layer K-Max pooling (k=3) Folding Wide convolution (m=2) Dynamic k-max pooling (k= f(s) =5) Wide convolution (m=3) Projected sentence matrix (s=7) game's the same, just got more fierce Deep learning should allow us to learn hierarchical generalisations.

  29. Deep Learning and NLP: Question Answer Selection When did James Dean die? Generalisation Generalisation In 1955, actor James Dean was killed in a two-car collision near Cholame, Calif. Beyond classification, deep models for embedding sentences have seen increasing success.

  30. Deep Learning and NLP: Question Answer Selection When did James Dean Die ? g In 1955, actor was killed in James Dean a two-car collision near Cholame, Calif. Recurrent neural networks provide a very practical tool for sentence embedding.

  31. 我 一 杯 白 葡萄酒 。 Deep Learning for NLP: Machine Translation i 'd like a glass of white wine , please . Generation Generalisation � � We can even view translation as encoding and decoding sentences.

  32. Deep Learning for NLP: Machine Translation Dogs love bones </s> Les chiens aiment les os ||| Dogs love bones Source sequence Target sequence Recurrent neural networks again perform surprisingly well.

  33. Supervised Reading Comprehension To achieve our aim of training supervised machine learning models for machine reading and comprehension, we must first find data.

  34. Supervised Reading Comprehension The CNN and DailyMail websites provide paraphrase summary sentences for each full news story.

  35. Supervised Reading Comprehension CNN article: Document The BBC producer allegedly struck by Jeremy Clarkson will not press charges against the “Top Gear” host, his lawyer said Friday. Clarkson, who hosted one of the most-watched television shows in the world, was dropped by the BBC Wednesday after an internal investigation by the British broadcaster found he had subjected producer Oisin Tymon “to an unprovoked physical and verbal attack.” . . . Query Producer X will not press charges against Jeremy Clarkson, his lawyer says. Answer Oisin Tymon We formulate Cloze style queries from the story paraphrases.

  36. Supervised Reading Comprehension From the Daily Mail: • The hi-tech bra that helps you beat breast X ; • Could Saccharin help beat X ?; • Can fish oils help fight prostate X ? An ngram language model would correctly predict ( X = cancer ), regardless of the document, simply because this is a frequently cured entity in the Daily Mail corpus.

  37. Supervised Reading Comprehension MNIST example generation: We generate quasi-synthetic examples from the original document-query pairs, obtaining exponentially more training examples by anonymising and permuting the mentioned entities.

  38. Supervised Reading Comprehension Original Version Anonymised Version Context The BBC producer allegedly struck by Jeremy the ent381 producer allegedly struck by ent212 will Clarkson will not press charges against the “Top not press charges against the “ ent153 ” host , his Gear” host, his lawyer said Friday. Clarkson, who lawyer said friday . ent212 , who hosted one of the hosted one of the most-watched television shows in most - watched television shows in the world , was the world, was dropped by the BBC Wednesday after dropped by the ent381 wednesday after an internal an internal investigation by the British broadcaster investigation by the ent180 broadcaster found he found he had subjected producer Oisin Tymon “to an had subjected producer ent193 “ to an unprovoked unprovoked physical and verbal attack.” . . . physical and verbal attack . ” . . . Query Producer X will not press charges against Jeremy producer X will not press charges against ent212 , his Clarkson, his lawyer says. lawyer says . Answer Oisin Tymon ent193 Original and anonymised version of a data point from the Daily Mail validation set. The anonymised entity markers are constantly permuted during training and testing. + Barun, Shantanu, Arindam, Ankit, Daraksha, Dinesh - Akshay/Barun : errors introduced by co-reference system?

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend