In the last chapter, you used the seq2seq model along with the Attention to perform a language translation. In this chapter, I will show you a more sophisticated technique for Natural Language Processing. You will learn to use the latest innovation in natural language modeling called Transformer. The Transformer model eliminates the need for LSTMs and produces far better results than the seq2seq model that uses LSTMs. So, let us understand what a Transformer model is.
In the last chapter, you saw the importance of Attention while ...