Natural Language Processing

Course 1: Introduction to Natural Language Processing

Introduction to Natural Language Processing : What is NLP ? What is it used for ?

Text Preprocessing : Preprocessing in Natural Language Processing (NLP) is the process by which we try to “standardize” the text we want to analyze.

Course 2: Core NLP Libraries

Introduction to NLTK: Todo

Introduction to SpaCy: Todo

Course 3: Text Embedding

Text Embedding with Bag-Of-Words and TF-IDF : In order to analyze text and run algorithms on it, we need to embed the text. The notion of embedding simply means that we’ll conver the input text into a set of numerical vectors that can be used into algorithms. In this article, we’ll cover BOW and TF-IDF, two simple techniques for embedding.

Text Embedding with Word2Vec : A deeper dive into the state of the art embedding technique : Word2Vec.

Data Augmentation in NLP : Details of the implementation of “Easy Data Augmentation” paper.

Course 4: Text Similarity

What is text similarity used for?: Todo

Similarity metrics and implementation: Todo

Course 5: Text Classification

Classical methods for text classification: Todo

Deep Learning Methods for text classification: Todo

Course 6: Sequence Models and Transformers

What are sequence models?: Todo

What are transformers?: Todo

Introduction to Hugging Face’s library: Todo

All codes and exercises are accessible on this repo. Don’t hesitate to show your suppot and star the repo: