Word Embedding
Praveen Krishnan
CVIT, IIIT Hyderabad
June 22, 2017
1
Word Embedding Praveen Krishnan CVIT, IIIT Hyderabad June 22, 2017 - - PowerPoint PPT Presentation
Word Embedding Praveen Krishnan CVIT, IIIT Hyderabad June 22, 2017 1 Outline Introduction Classical Methods Language Modeling Neural Language Model Challenges of SoftMax Hierarchical Softmax Margin Based Hinge Loss Sampling Based
1
2
3
4
5
6
7
8
9
10
11
12
12
13
14
◮ Hierarchical Softmax ◮ Differentiated Softmax ◮ CNN-Softmax
◮ Importance Sampling. ◮ Margin based Hinge loss. ◮ Noise Contrastive Approximation ◮ Negative Sampling ◮ ... 15
16
◮ Morin and Bengio: Using synsets in WordNet as clusters of tree. ◮ Mikolov et. al.: Use Huffman tree which takes into account the
17
18
19
20
◮ Removed hidden layer. ◮ Use of additional context for training LM’s. ◮ Introduced newer training strategies using huge database of words
21
22
23
24
◮ Easy to sample. ◮ Allows analytical expression to log pdf. ◮ Close to actual data distribution. E.g. Uniform or empirical
25
26
27
w)
w)
w) + k Q(w)
wi)
wi) + k Q(wi)+ k
˜ wij)
˜ wij) + k Q( ˜
28
29
30
31
32