Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Evolution of word Embedding in NLP #24

Open
1 of 3 tasks
Praveenrajan27 opened this issue Oct 22, 2019 · 4 comments
Open
1 of 3 tasks

Evolution of word Embedding in NLP #24

Praveenrajan27 opened this issue Oct 22, 2019 · 4 comments
Assignees
Labels

Comments

@Praveenrajan27
Copy link

Title

Evolution of word Embedding in NLP

Description

Modeling natural language is a really complex problem because natural language doesn’t follow any mathematical construct.
It is very abstract which makes it very hard to train language. In past there have been rule based solutions but it had limited performance in terms of automation
Over time it became evident that if we want to solve NLP problems efficiently we have to treat it as a mathematical problem and from there started the journey of embeddings
From representing text in a simple one hot encoded method to more advanced representation like vectors notation, we have come a long way.
This talk takes a journey through how this representation known as the embedding have evolved over the years.

Duration

  • 30 min
  • 45 min

Audience

Basic understanding of NLP would be good. This talk is for Beginner to Intermediate level

Outline

A detailed outline for your talk. The more detailed the better. (1000 words)

  • Intro -

  • Text analytics and how it is relevant today

  • Challenges in working with NLP data - need to represent text as mathematical representation Why do need to represent text in mathematical format

  • Early stage - representing text in mathematical space
    1. Early forms of representation using one hot encoding
    2. Measuring text relevancy using TF-IDF and how it improves representation quality
    3. Practical applications of word embedding leveraging TF-IDF

  • Mid Stage - capturing relationship in words
    1. Intro of word2vec model
    Architectural overview - Skip gram and Continuous bag of words
    2. How word2vec started the legacy of pretrained models
    3. Limitations of word2vec model

  • Current - Capturing contextual information in embeddings
    Intro to attention mechanism
    1. Overview of Attention based model - Bi,BERT
    2. Transfer Learning in NLP - ImageNet moment
    3. Architectural comparison between current state of the art NLP models - BERT, XLNet and GPT 1,2
    Applications leveraging SOTA models

Additional notes

I am Data science Professional with VMware with 6+years of exp. I have completed my masters in software Engineering. My interests are in NLP and Language models


  • Don't record this talk.

Check this if you don't want your talk to be recorded.

@vinayak-mehta
Copy link
Member

@Praveenrajan27 Are you available to present this talk on Nov 9?

@Praveenrajan27
Copy link
Author

Yes I am!

@TrigonaMinima
Copy link

Hi @Praveenrajan27 will you be available for the talk on 29th Feb?

@Praveenrajan27
Copy link
Author

Praveenrajan27 commented Feb 28, 2020 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants