neural classification of linguistic coherence using long
play

Neural Classification of Linguistic Coherence using Long Short-Term - PowerPoint PPT Presentation

Neural Classification of Linguistic Coherence using Long Short-Term Memories Pashutan Modaresi, Matthias Liebeck and Stefan Conrad Order of Sentences Hi! My name is Alan. My name is Alan. Hi! Is what makes a text semantically I am a


  1. Neural Classification of Linguistic Coherence using Long Short-Term Memories Pashutan Modaresi, Matthias Liebeck and Stefan Conrad

  2. Order of Sentences Hi! My name is Alan. My name is Alan. Hi! Is what makes a text semantically I am a computer scientist! I am a computer scientist! meaningful Hi! ➔ Hi! I am a computer scientist! I am a computer scientist! My name is Alan. ➔ My name is Alan. My name is Alan. Hi! I am a computer scientist! ➔ I am a computer scientist! My name is Alan. Hi! I am a computer scientist! My name is Alan. Hi!

  3. Humans vs. Machines Discourse Coherence Linguistic Contradiction Question Is there a need to teach Linguistic Redundancy all these abilities to a machine? Pragmatics

  4. 0, 1 Sentence Ordering -1, 0, 1 Question? Hi! My name is Alan. What about the sizes of m and m’ ? Should they be equal?

  5. Many Applications! Focus was TEXT SUMMARIZATION in the news domain Question? What are the other applications of sentence ordering?

  6. Treat the problem as a classification task Number of Instances Question Why do we use the negative log-likelihood and not the log-likelihood? Class probability of the n-th pair

  7. Deep Neural Architecture S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  8. Deep Neural Architecture One-Hot Encoding S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  9. Tip Embedding: Simple matrix multiplication Deep Neural Architecture with input vector Init the matrix E Embedding S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  10. Tip Concatenate the embeddings Deep Neural Architecture Merge S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  11. Tip LSTM: Just a special kind of RNNs addressing Deep Neural Architecture their difficulties Long Short-Term Memory S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  12. Tip Dropout: Sets a random set of its arguments to Deep Neural Architecture zero. Regularization S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  13. Tip Dropout: Sets a random set of its arguments to Deep Neural Architecture zero. Softmax S1 +1 0 LSTM Dropout LSTM Dropout -1 S2

  14. Data How to collect the required data to train the -1 network? +1 Binary ➔ Correct Wrong Order Order Ternary ➔ +1 -1 0 Correct Wrong Mising Order Order Context

  15. Baseline - SVM English German Binary Ternary Binary Ternary 0.24 0.16 0.25 0.16 SVMs: Not really appropriate for sequential modelling

  16. Macro-Averaged F1

  17. Lessons Learned ● Use appropriate tools for sequence modeling ● RNNs are slow. First train on a subset of data ● Train deep models with lots of data points ● Find a way to automatically annotate data ● Use regularization (be generous)

  18. Thank You For Your Attention

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend