Unit - 3 Distributional Semantics and Word Embedding
Unit - 3 Distributional Semantics and Word Embedding
Word
Embeddings:
Intuition
We want a model which will be able to state that coffee and tea are
close and they are also close with words like cup, caffeine, drink,
sugar etc.
• All the available algorithms are based on the following
principles:
• Semantically similar words are mapped to nearby points.
• The basic idea is the Distributional Hypothesis: words
that appear in the same contexts share semantic meaning
Word like tea and coffee.
• The most common algorithms are the Word2Vec (Mikolov et
Embeddings: al. (2013) at Google) and GloVe (2014 Stanford) where they
take as input a large corpus of text and produce a vector
The space typically of 100-300 dimensions.
• So the corresponding Word Embeddings of the words coffee,
Algorithms tea and laptop would look like
Language
Construct
Semantics vs
pragmatics
Study of meaning of
words,sentences and phrases
Pharasal/Sentential semantics-
Syntactic units larger thana
word.
Semantics and why??
Semantics is the study of meaning.
What is meaning?
How can the meanings of words combine to form the meaning of sentences?
Do two people mean the same thing when they utter the word ‘cat’.