Building a language model with a gated recurrent unit

To demonstrate the flexibility of recurrent networks, we are going to do something different in the final section of this chapter. Up until now, we have been working with probably the most-used testing data set, MNIST. This dataset has characteristics that are well known and it is extremely useful for comparing different types of models and testing different architectures and parameter sets. However, there are some tasks, such as natural language processing, that quite obviously require an entirely different type of dataset.

Also, the models we have built so far have been focused on one simple task: classification. This is the most straightforward machine learning task. To give you a flavor ...

Get Deep Learning with PyTorch Quick Start Guide now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.