#TagSpace: Semantic Embeddings from Hashtags Jason Weston, Sumit - - PowerPoint PPT Presentation
#TagSpace: Semantic Embeddings from Hashtags Jason Weston, Sumit - - PowerPoint PPT Presentation
#TagSpace: Semantic Embeddings from Hashtags Jason Weston, Sumit Chopra, Keith Adams 2014 Jack Lanchantin Motivation Word and document embeddings are difficult to learn Most current techniques use unsupervised methods word2vec
Motivation
- Word and document embeddings are difficult to learn
- Most current techniques use unsupervised methods
- word2vec learns word embeddings by trying to
predict each word in a doc based on surrounding text
- Hashtags: labels of text for such as sentiment (#happy)
- r topic annotation (#nyc) written by the post author
- Hashtag prediction provides better way to learn word
and document embeddings than unsupervised learning because hashtags provide stronger semantic guidance
Overview Model Hashtag Prediction Document Recommendation Conclusion
Overview
- #TagSpace: Convolutional Neural Network that
learns features (embeddings) of short textual posts using hashtags as the supervised signal
- Train the network to be able to optimally
predict hashtags on test posts
- The learned embedding of text (ignoring the
hashtag labels) is useful for other tasks such as document recommendation
Overview Model Hashtag Prediction Document Recommendation Conclusion
Neural Net For Scoring a (doc, hashtag) Pair
Assigning a d dimensional vector to each of the l words Hidden network layers Representation of entire document Assigning a d dimensional vector to the hashtag Scoring function
- Given a document, rank all hashtags by score:
- Loss function is used to approximately optimize
the top of the ranked list – useful for P and R@k
- More energy spent on improving ranking of
positive labels near the top of ranked list
Training the Scoring Function
Overview Model Hashtag Prediction Document Recommendation Conclusion
Hashtag Prediction
- Goal: Rank a post’s ground truth hashtags higher
than hashtags it does not contain
- Test using: Precision @ 1, Recall @10, mean rank
for the hashtags of 50,000 test posts
- Compared to 4 other models:
–Frequency: always ranks hashtags by training frequency –#words: “crazy commute this am” → #crazy, #commute, #this, #am –Word2vec (unsupervised) –WSABIE (supervised)
Data
Business, Celebrity, Brand, or Product Individual users
#TagSpace Examples (256 dim)
Post Predicted Hashtags
Hashtag Prediction Results
Overview Model Hashtag Prediction Document Recommendation Conclusion
Personalized Document Recommendation
- Goal: extend the learned representations from
predicting hashtags to do other tasks
- Document recommendation: recommending
documents to users based on interaction history
- Used day-long interaction histories for 34,000
people on Facebook
- Text of posts that he/she liked, clicked, replied to
- Given n-1 trailing posts, predict the nth post by
ranking it against 10,000 other posts
- Score of nth post is obtained by max embedding
similarity over n-1 posts
- Used cosine similarity between post embeddings
Document Recommendation Results
TF-IDF weighted bag of words baseline Best results come from summing BOW scores w/ Tagspace
Overview Model Hashtag Prediction Document Recommendation Conclusion
Conclusion
- Outperformed all comparison models in hashtag
prediction
- Model scales very well when considering a large
number (millions) of hashtags
- Logistic regression and SVMs do not
- Semantics of hashtags cause #TagSpace to learn
features that capture important aspects of text
- Able to port the learned embeddings to the task
- f personalized document recommendation
with better accuracy than other models
#TagSpace: https://research.facebook.com/publications/279494668926031/- tagspace-semantic-embeddings-from-hashtags/ WSABIE: http://www.thespermwhale.com/jaseweston/papers/wsabie-ijcai.pdf Word2Vec: http://arxiv.org/abs/1301.3781