Training an NMT jointly with word embeddings

Here we will discuss how we can train an NMT jointly with word embeddings. We will be covering two concepts in this section:

  • Training an NMT jointly with a word embedding layer
  • Using pretrained embeddings instead of randomly initializing the embeddings layer

There are several multilingual word embedding repositories available:

From these, we will use the CMU embeddings (~200 MB) as it's much smaller compared with fastText (~5 GB). We first need to download the German (multilingual_embeddings.de) and English (multilingual_embeddings.en ...

Get Natural Language Processing with TensorFlow now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.