SlideShare a Scribd company logo
Neural Word
Embedding And
Language Modelling
A Survey Paper Presentation by
Riddhi Jain
Graduate Student Computer Engineering, SJSU
(014600716)
Paper:
A survey on Neural Word Embeddings
Authors:
ERHAN SEZERER
SELMA TEKIR
Article Link: https://arxiv.org/pdf/2110.01804.pdf
Published Date: 2021
About
What is VSM?
The vector space model is an algebraic model that represents objects (like text)
as vectors. This makes it easy to determine the similarity between words or the
relevance between a search query and document. Cosine similarity is often used
to determine similarity between vectors.
Neural Word Embeddings
Neural network architecture is constructed to predict the next word given the set
of neighboring words in the sequence in neural language modeling.
Word Embeddings with Improved Language
Models
● Early Word Embeddings
● Embeddings Target Specific Semantic Relations
● Sense Embeddings
● Morpheme Embeddings
Early Word Embeddings
Word2vec is the first neural word embedding model that efficiently computes
representations to leverage the context of target words
Two word2vec variants:
● CBOW (Continuous Bag of words)
○ Example - “nature is pleased with simplicity”
● Skip-gram
Embeddings Target Specific Semantic
Relations
Example - “She took a sip of hot coffee” and “He is taking a sip of cold water”
Types of algorithms
- SGNS
- GloVe
- ATTRACT and REPEL
Sense Embeddings
● Early word embeddings unite all the senses of a word into one
representation.
● In reality, a word gets meaning in its use and can mean different things in
varying contexts.
● When the issue becomes labeling those sense groups, the task becomes a
supervised one
Morpheme Embeddings
Proposes several ways to target morphological information in order to obtain
sub-word information for solving the rare/unknown word problem of earlier word
embedding methods and also to have better representations of words for
morphologically rich languages.
Two ways
- Training Morphological Embeddings from Scratch
- Adjusting the Existing Embeddings
Datasets
● Similarity Tasks
● Analogy Task
Google Analogy Task with 8869 semantic and 10675 syntactic
● Synonym Selection Tasks
● Downstream Tasks
GLUE benchmark dataset
Human-level language understanding is
one of the oldest challenges in computer
science. Pre-trained language models’
knowledge has been transferred to fine-
tuned task-specific models. Multi-modal
language models are based on human
language acquisition, where learning starts
with concrete concepts through images
early on.
Conclusion

More Related Content

PPT
Visual thinking colin_ware_lectures_2013_12_pictures and words
PDF
A supervised word sense disambiguation method using ontology and context know...
PPT
Connectionism pp 3
PPTX
Possible Word Representation
PDF
A Survey on Word Sense Disambiguation
PDF
Comprehension and Language
PDF
International Journal of Engineering and Science Invention (IJESI)
PDF
Recurrent Convolutional Neural Networks for Text Classification
Visual thinking colin_ware_lectures_2013_12_pictures and words
A supervised word sense disambiguation method using ontology and context know...
Connectionism pp 3
Possible Word Representation
A Survey on Word Sense Disambiguation
Comprehension and Language
International Journal of Engineering and Science Invention (IJESI)
Recurrent Convolutional Neural Networks for Text Classification

What's hot (20)

PDF
Improvement wsd dictionary using annotated corpus and testing it with simplif...
PDF
Verbal cognition: vector space analysis by Chuluundorj.B /University of the H...
PPTX
Representation of syntax, semantics and Predicate logics
PPTX
Object oriented programming mine normalized
PDF
GDG Tbilisi 2017. Word Embedding Libraries Overview: Word2Vec and fastText
PPTX
Connectionist and Dynamical Systems approach to Cognition
PPT
Emergentism Presentation
PDF
Word Sense Disambiguation and Induction
PPTX
Bridging the gap between AI and UI - DSI Vienna - full version
PPT
Using construction grammar in conversational systems
PDF
Review: Deep contextualized word representations
PPT
Which Rationality For Pragmatics6
PPTX
Construction Grammar
DOC
Doc format.
PPTX
Intent Classifier with Facebook fastText
PDF
Making sense of word senses: An introduction to word-sense disambiguation and...
PDF
Text-Analysis-Orange.pdf
PDF
AN EMPIRICAL STUDY OF WORD SENSE DISAMBIGUATION
PPTX
Icwl2015 wahl
PDF
Semantic Rules Representation in Controlled Natural Language in FluentEditor
Improvement wsd dictionary using annotated corpus and testing it with simplif...
Verbal cognition: vector space analysis by Chuluundorj.B /University of the H...
Representation of syntax, semantics and Predicate logics
Object oriented programming mine normalized
GDG Tbilisi 2017. Word Embedding Libraries Overview: Word2Vec and fastText
Connectionist and Dynamical Systems approach to Cognition
Emergentism Presentation
Word Sense Disambiguation and Induction
Bridging the gap between AI and UI - DSI Vienna - full version
Using construction grammar in conversational systems
Review: Deep contextualized word representations
Which Rationality For Pragmatics6
Construction Grammar
Doc format.
Intent Classifier with Facebook fastText
Making sense of word senses: An introduction to word-sense disambiguation and...
Text-Analysis-Orange.pdf
AN EMPIRICAL STUDY OF WORD SENSE DISAMBIGUATION
Icwl2015 wahl
Semantic Rules Representation in Controlled Natural Language in FluentEditor
Ad

Similar to Neural word embedding and language modelling (20)

PDF
Challenges in transfer learning in nlp
PDF
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
PDF
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
PDF
Continuous bag of words cbow word2vec word embedding work .pdf
DOCX
A neural probabilistic language model
PDF
Unsupervised hindi word sense disambiguation using graph based centrality mea...
PDF
Generative Artificial Intelligence and Large Language Model
PDF
Automatic classification of bengali sentences based on sense definitions pres...
PDF
Effect of word embedding vector dimensionality on sentiment analysis through ...
DOCX
Natural Language Processing an introduction
PDF
AN APPROACH TO WORD SENSE DISAMBIGUATION COMBINING MODIFIED LESK AND BAG-OF-W...
PDF
An approach to word sense disambiguation combining modified lesk and bag of w...
PDF
[Paper Reading] Unsupervised Learning of Sentence Embeddings using Compositi...
PDF
IRJET- Survey on Deep Learning Approaches for Phrase Structure Identification...
PPTX
PPT Unit 5=software- engineering-21.pptx
PDF
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
PDF
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
PDF
CHUNKER BASED SENTIMENT ANALYSIS AND TENSE CLASSIFICATION FOR NEPALI TEXT
PPTX
short_story.pptx
PPTX
Text Mining for Lexicography
Challenges in transfer learning in nlp
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
Continuous bag of words cbow word2vec word embedding work .pdf
A neural probabilistic language model
Unsupervised hindi word sense disambiguation using graph based centrality mea...
Generative Artificial Intelligence and Large Language Model
Automatic classification of bengali sentences based on sense definitions pres...
Effect of word embedding vector dimensionality on sentiment analysis through ...
Natural Language Processing an introduction
AN APPROACH TO WORD SENSE DISAMBIGUATION COMBINING MODIFIED LESK AND BAG-OF-W...
An approach to word sense disambiguation combining modified lesk and bag of w...
[Paper Reading] Unsupervised Learning of Sentence Embeddings using Compositi...
IRJET- Survey on Deep Learning Approaches for Phrase Structure Identification...
PPT Unit 5=software- engineering-21.pptx
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
CHUNKER BASED SENTIMENT ANALYSIS AND TENSE CLASSIFICATION FOR NEPALI TEXT
short_story.pptx
Text Mining for Lexicography
Ad

Recently uploaded (20)

PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
PDF
Pre independence Education in Inndia.pdf
PDF
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PDF
Classroom Observation Tools for Teachers
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PDF
Microbial disease of the cardiovascular and lymphatic systems
PPTX
PPH.pptx obstetrics and gynecology in nursing
PDF
Chapter 2 Heredity, Prenatal Development, and Birth.pdf
PPTX
Cell Types and Its function , kingdom of life
PDF
RMMM.pdf make it easy to upload and study
PPTX
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
PDF
O5-L3 Freight Transport Ops (International) V1.pdf
PPTX
Microbial diseases, their pathogenesis and prophylaxis
PDF
Basic Mud Logging Guide for educational purpose
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PPTX
master seminar digital applications in india
PPTX
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
STATICS OF THE RIGID BODIES Hibbelers.pdf
Pre independence Education in Inndia.pdf
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
Supply Chain Operations Speaking Notes -ICLT Program
Classroom Observation Tools for Teachers
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
Microbial disease of the cardiovascular and lymphatic systems
PPH.pptx obstetrics and gynecology in nursing
Chapter 2 Heredity, Prenatal Development, and Birth.pdf
Cell Types and Its function , kingdom of life
RMMM.pdf make it easy to upload and study
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
O5-L3 Freight Transport Ops (International) V1.pdf
Microbial diseases, their pathogenesis and prophylaxis
Basic Mud Logging Guide for educational purpose
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
master seminar digital applications in india
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx

Neural word embedding and language modelling

  • 1. Neural Word Embedding And Language Modelling A Survey Paper Presentation by Riddhi Jain Graduate Student Computer Engineering, SJSU (014600716)
  • 2. Paper: A survey on Neural Word Embeddings Authors: ERHAN SEZERER SELMA TEKIR Article Link: https://arxiv.org/pdf/2110.01804.pdf Published Date: 2021 About
  • 3. What is VSM? The vector space model is an algebraic model that represents objects (like text) as vectors. This makes it easy to determine the similarity between words or the relevance between a search query and document. Cosine similarity is often used to determine similarity between vectors.
  • 4. Neural Word Embeddings Neural network architecture is constructed to predict the next word given the set of neighboring words in the sequence in neural language modeling.
  • 5. Word Embeddings with Improved Language Models ● Early Word Embeddings ● Embeddings Target Specific Semantic Relations ● Sense Embeddings ● Morpheme Embeddings
  • 6. Early Word Embeddings Word2vec is the first neural word embedding model that efficiently computes representations to leverage the context of target words Two word2vec variants: ● CBOW (Continuous Bag of words) ○ Example - “nature is pleased with simplicity” ● Skip-gram
  • 7. Embeddings Target Specific Semantic Relations Example - “She took a sip of hot coffee” and “He is taking a sip of cold water” Types of algorithms - SGNS - GloVe - ATTRACT and REPEL
  • 8. Sense Embeddings ● Early word embeddings unite all the senses of a word into one representation. ● In reality, a word gets meaning in its use and can mean different things in varying contexts. ● When the issue becomes labeling those sense groups, the task becomes a supervised one
  • 9. Morpheme Embeddings Proposes several ways to target morphological information in order to obtain sub-word information for solving the rare/unknown word problem of earlier word embedding methods and also to have better representations of words for morphologically rich languages. Two ways - Training Morphological Embeddings from Scratch - Adjusting the Existing Embeddings
  • 10. Datasets ● Similarity Tasks ● Analogy Task Google Analogy Task with 8869 semantic and 10675 syntactic ● Synonym Selection Tasks ● Downstream Tasks GLUE benchmark dataset
  • 11. Human-level language understanding is one of the oldest challenges in computer science. Pre-trained language models’ knowledge has been transferred to fine- tuned task-specific models. Multi-modal language models are based on human language acquisition, where learning starts with concrete concepts through images early on. Conclusion

Editor's Notes

  • #2: Hello
  • #4: Before neural representation learning, representations of words or documents have been computed using the vector space model. In VSM [120], frequencies of words in documents are considered to form a term-document matrix. Although these count-based representations are proved helpful in addressing semantics, they are the bag of words approaches and are not able to capture both syntactical and semantic features at the same time, which is required for performing well in NLP tasks.
  • #5: Neural network architecture is constructed to predict the next word given the set of neighboring words in the sequence in neural language modeling. In the iterative processing of this prediction over a large corpus, the learned weights in the hidden layers serve as neural embeddings for words.
  • #6: Although the initial word embedding models successfully identified semantic and syntactic similarities of words, they still need to be improved to address specific semantic relations among words. consider the sentences "She took a sip of hot coffee" and "He is taking a sip of cold water." The antonyms "cold" and "hot" are deemed to be similar since their context is similar. Therefore, it becomes an issue to differentiate the synonyms "warm" and "hot" from the antonyms "cold" and "hot" considering they have similar contexts in most occurrences. Early word embeddings unite all the senses of a word into one representation. In reality, a word gets meaning in its use and can mean different things in varying context. Sense Embeddings are used to get word sense discrimination as the decomposition of a word’s occurrences into same sense groups
  • #7: Word2vec is the first neural word embedding model that efficiently computes representations to leverage the context of target words. It can be considered as the initiator of early word embeddings. Word2vec has two variants: Continuous bag of words model (CBOW) and Skip-gram model. In CBOW - middle word is predicted given the context, set of neighbouring left and right. When the input sentence "nature is pleased with simplicity" is processed, the system predicts the middle word "pleased" given the left and right context In Skip-gram system predicts the most probable context words for a given input word. In terms of a language model, while CBOW predicts an individual word’s probability, Skip-gram outputs the probabilities of a set of words, defined by a given context size.
  • #8: Although the initial word embedding models successfully identified semantic and syntactic similarities of words, they still need to be improved to address specific semantic relations among words. consider the sentences "She took a sip of hot coffee" and "He is taking a sip of cold water." The antonyms "cold" and "hot" are deemed to be similar since their context is similar. Therefore, it becomes an issue to differentiate the synonyms "warm" and "hot" from the antonyms "cold" and "hot" considering they have similar contexts in most occurrences.
  • #9: Early word embeddings unite all the senses of a word into one representation. In reality, a word gets meaning in its use and can mean different things in varying context. Sense Embeddings are used to get word sense discrimination as the decomposition of a word’s occurrences into same sense groups
  • #10: The quest for morphological representations is a result of two important limitations of earlier word embedding models. The first point is, words are not the smallest units of meaning in language. Even if a model does not see the word unpleasant in the training it should be able to deduce that it is the negative form of pleasant. Word embedding methods that don’t take morphological information into account can not produce any results in such a situation. The second limitation is the data scarcity problem of morphologically rich languages and agglutinative languages. Unlike English, morphologically rich languages have many more noun and/or verb forms inflected by gender, case, or number, which may not exist in the training corpora. The same thing is also valid for agglutinative languages in which words can have many forms according to the suffix(es) they take. Therefore, models that take morphemes/lexemes into account needed.