10 Sequence-to-sequence models and attention

This chapter covers

  • Mapping one text sequence to another with a neural network
  • Understanding sequence-to-sequence tasks and how they’re different from the others you’ve learned about
  • Using encoder-decoder model architectures for translation and chat
  • Training a model to pay attention to what is important in a sequence

You now know how to create natural language models and use them for everything from sentiment classification to generating novel text (see chapter 9).

Could a neural network translate from English to German? Or even better, would it be possible to predict disease by translating genotype to phenotype (genes to body type)?[1] And what about the chatbot we’ve been talking about since the ...

Get Natural Language Processing in Action now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.