4. Deep Learning for Text – Embeddings

Overview

In this chapter, we will begin our foray into Natural Language Processing for text. We will start by using the Natural Language Toolkit to perform text preprocessing on raw text data, where we will tokenize the raw text and remove punctuations and stop words. As we progress through this chapter, we will implement classical approaches to text representation, such as one-hot encoding and the TF-lDF approach. This chapter demonstrates the power of word embeddings and explains the popular deep learning-based approaches for embeddings. We will use the Skip-gram and Continuous Bag of Words algorithms to generate our own word embeddings. We will explore the properties of the embeddings, the different ...

Get The Deep Learning Workshop now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.