CMSC - 676 Classifications using word embedding techniques
Presented By - Prachi Bhalerao (prachib1@umbc.edu)
CMSC - 676 Classifications using word embedding techniques - - PowerPoint PPT Presentation
CMSC - 676 Classifications using word embedding techniques Presented By - Prachi Bhalerao (prachib1@umbc.edu) Introduction Need - Finding similarity between words and extracting semantic/contextual information as much as possible is viable.
Presented By - Prachi Bhalerao (prachib1@umbc.edu)
Predicts target word from given context
Predicts the context from given word
topics is proposed.
under a topic.
topic embeddings as the weight value, the word embeddings
embeddings and enhance the semantics of the word embeddings.
sentence embeddings, which is the characteristic value of the blog
task of named entity recognition (NER).
belong to which predefined category.
Linear Support Vector algorithm.
Results-
the unlabelled data set but only to a limit (around 300 000 types, in this paper) at which it even started to drop.
Measures given are achieved from adding clusters at granularity 1000, built from word2vec models trained on the various data sets, to the NER classifier. One possible explanation for the stagnating performance of the larger data set is that
for optimal training (e.g. higher vector dimensionality or more training iterations).
Average of the contexts of all the words with same meaning is taken. When put into practice, this can significantly impact on the performance of ML systems posing a potential problem for conversational agents and text classifiers e.g. Apple, like
Inflected forms (past tense or participle, for example) of verbs. That’s because some word inflections appear less frequently than others in certain contexts -> Fewer examples of those ‘less common’ words in context for the algorithm to learn from them -> ‘less similar’ vectors
Phys.: Conf. Ser. 1168 052004; https://iopscience.iop.org/article/10.1088/1742- 6596/1168/5/052004/pdf
https://www.ep.liu.se/ecp/109/030/ecp15109030.pdf
Embeddings’ ACL 2019; https://arxiv.org/abs/1906.09821
acm; https://dl.acm.org/doi/abs/10.1145/775047.775076