Sequence Prediction Using Neural Network Classifiers
Yanpeng Zhao ShanghaiTech University
ICGI, Oct 7th, 2016, Delft, the Netherlands
Sequence Prediction Using Neural Network Classifiers Yanpeng Zhao - - PowerPoint PPT Presentation
Sequence Prediction Using Neural Network Classifiers Yanpeng Zhao ShanghaiTech University ICGI, Oct 7 th , 2016, Delft, the Netherlands Sequence Prediction Whats the next symbol? 4 3 5 0 4 6 1 3 1 ? Classification Perspective -1
ICGI, Oct 7th, 2016, Delft, the Netherlands
What’s the next symbol?
Input Sequence
Multinomial the most likely next symbol
King – Man + Woman ≈ Queen
Images are from: https://blog.acolyer.org/2016/04/21/the-amazing-power-of-word-vectors/
5 Label
Word vectors are concatenated or stacked Predict the next symbol from the previous = 15 symbols, each represented by a 30-dimension vector
Input Test Sequence
Multinomial the most likely next symbol
+1
+1
Input Hidden Layer 1 Hidden Layer 2 Softmax Output
= 1 a = = 1 a = = 1 =
+1
+1
Input Hidden Layer 1 Hidden Layer 2 Softmax Output
= 1 a = = 1 a = = 1 =
|| = 450 : 15 symbols with a 30-dimension vector for each symbol = 750 and = 1000
CNN model architecture adapted from Yoon Kim. Convolutional neural networks for sentence classication. arXiv preprint arXiv:1408.5882, 2014
k = 15, = 30 Filter windows (height) of 10, 11, 12, 13, 14, 15 ; 200 feature maps for each window
ℎ
=
⊗ + ⊗
() Images are from: http://colah.github.io/posts/2015-08-Understanding-LSTMs/
Time step is 15, and ℎ of dim 32 is fed to a logistic regression classifier
Label
We set to 2, 3, 4, 5, 6 with weights 0.3, 0.2, 0.2, 0.15, 0.15 respectively
Total score on private test sets is 10.160324
prediction
8.666 7.444 9.802 9.593 9.237 9.325
5 10 15 3-Gram SL MLP CNN WnGram LSTM
Total scores by different models