top down tree long short term memory networks
play

Top-down Tree Long Short-Term Memory Networks Xingxing Zhang , Liang - PowerPoint PPT Presentation

Top-down Tree Long Short-Term Memory Networks Xingxing Zhang , Liang Lu, Mirella Lapata School of Informatics, University of Edinburgh 12th June, 2016 Zhang et al., 2016 Tree LSTM 12th June, 2016 1 / 18 Sequential Language Models n P (


  1. Top-down Tree Long Short-Term Memory Networks Xingxing Zhang , Liang Lu, Mirella Lapata School of Informatics, University of Edinburgh 12th June, 2016 Zhang et al., 2016 Tree LSTM 12th June, 2016 1 / 18

  2. Sequential Language Models n � P ( S = w 1 , w 2 , . . . , w n ) = P ( w i | w 1: i − 1 ) (1) i =1 State of the Art based on Long Short Term Memory Network Language Model (Hochreiter and Schmidhuber, 1997; Sundermeyer et al., 2012) Billion word benchmark results reported in Jozefowicz et al., (2016) Models PPL KN5 67.6 LSTM 30.6 LSTM+CNN INPUTS 30.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 2 / 18

  3. Will tree structures help LMs? Zhang et al., 2016 Tree LSTM 12th June, 2016 3 / 18

  4. Will tree structures help LMs? Probably yes LMs based on Constituency Parsing (Chelba and Jelinek, 2000; Roark, 2001; Charniak, 2001) LMs based on Dependency Parsing (Shen et al., 2008; Zhang, 2009; Sennrich, 2015) Zhang et al., 2016 Tree LSTM 12th June, 2016 3 / 18

  5. LSTMs + Dependency Trees = TreeLSTMs + Why? Sentence Length N v.s. Tree Height log ( N ) Zhang et al., 2016 Tree LSTM 12th June, 2016 4 / 18

  6. LSTMs + Dependency Trees = TreeLSTMs + Why? Sentence Length N v.s. Tree Height log ( N ) How? Top-down Generation Breadth-first search reminiscent of Eisner (1996) Zhang et al., 2016 Tree LSTM 12th June, 2016 4 / 18

  7. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  8. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  9. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  10. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  11. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  12. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  13. Generation Process (Unlabeled Trees) The luxury auto manufacturer last year sold 1,214 cars in the U.S. Zhang et al., 2016 Tree LSTM 12th June, 2016 5 / 18

  14. Tree LSTM n � P ( S ) = P ( w i | w 1: i − 1 ) (2) i =1 ⇓ � P ( S | T ) = P ( w |D ( w )) (3) w ∈ BFS( T ) \ root D ( w ) is the Dependency Path of w . D ( w ) is a generated sub-tree. Works on projective and unlabeled dependency trees. Zhang et al., 2016 Tree LSTM 12th June, 2016 6 / 18

  15. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  16. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  17. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  18. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  19. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  20. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  21. Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 7 / 18

  22. One Limitation of Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 8 / 18

  23. Left Dependent Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 9 / 18

  24. Left Dependent Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 9 / 18

  25. Left Dependent Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 9 / 18

  26. Left Dependent Tree LSTM Zhang et al., 2016 Tree LSTM 12th June, 2016 9 / 18

  27. Experiments Zhang et al., 2016 Tree LSTM 12th June, 2016 10 / 18

  28. MSR Sentence Completion Challenge Training set: 49 million words (around 2 million sentences) development set: 4000 sentences test set: 1040 completion questions. Zhang et al., 2016 Tree LSTM 12th June, 2016 11 / 18

  29. Zhang et al., 2016 Tree LSTM 12th June, 2016 12 / 18

  30. Zhang et al., 2016 Tree LSTM 12th June, 2016 12 / 18

  31. Zhang et al., 2016 Tree LSTM 12th June, 2016 12 / 18

  32. Zhang et al., 2016 Tree LSTM 12th June, 2016 12 / 18

  33. Dependency Parsing Reranking Rerank 2nd Order MSTParser (McDonald and Pereira, 2006) We train TreeLSTM and LdTreeLSTM as language models. We only use words as input features; POS tags, dependency labels or composition features are not used. Zhang et al., 2016 Tree LSTM 12th June, 2016 13 / 18

  34. Dependency Parsing Reranking NN: Chen & Manning, 2014; S-LSTM: Dyer et al., 2015 Zhang et al., 2016 Tree LSTM 12th June, 2016 14 / 18

  35. Dependency Parsing Reranking NN: Chen & Manning, 2014; S-LSTM: Dyer et al., 2015 Zhang et al., 2016 Tree LSTM 12th June, 2016 14 / 18

  36. Dependency Parsing Reranking NN: Chen & Manning, 2014; S-LSTM: Dyer et al., 2015 Zhang et al., 2016 Tree LSTM 12th June, 2016 14 / 18

  37. Dependency Parsing Reranking NN: Chen & Manning, 2014; S-LSTM: Dyer et al., 2015 Zhang et al., 2016 Tree LSTM 12th June, 2016 14 / 18

  38. Tree Generation Four binary classifiers: Add Left? No! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  39. Tree Generation Four binary classifiers: Add Right? Yes! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  40. Tree Generation Four binary classifiers: Add Right? Yes! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  41. Tree Generation Four binary classifiers: Add Next Right? No! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  42. Tree Generation Four binary classifiers: Add Left? Yes! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  43. Tree Generation Four binary classifiers: Add Left? Yes! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  44. Tree Generation Four binary classifiers: Add Next Left? No! Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  45. Tree Generation Four binary classifiers: Add Left? Add Right? Add Next Left? Add Next Right? Features: hidden states and word embeddings Classifiers Accuracies Add-Left 94.3 Add-Right 92.6 Add-Nx-Left 93.4 Add-Nx-Right 96.0 Zhang et al., 2016 Tree LSTM 12th June, 2016 15 / 18

  46. Tree Generation Zhang et al., 2016 Tree LSTM 12th June, 2016 16 / 18

  47. Conclusions Syntax can help language modeling. Predicting tree structures with Neural Networks is possible. Next Steps: Sequence to Tree Models Tree to Tree Models code available: https://github.com/XingxingZhang/td-treelstm Thanks & Questions? Zhang et al., 2016 Tree LSTM 12th June, 2016 17 / 18

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend