April 2026
461 pages
17h 56m
English
A highlight in the 1990s was the consideration of time dependencies in the data, such as in the case of spoken language. RNNs, which have been known for some time, were used as the basis for this.
In 1997, Sepp Hochreiter and Jürgen Schmidhuber published an article that proposed a new architecture for networks that have a long short-term memory (LSTM). There are approaches (as you’ll see, for example, with regard to Q-learning in Chapter 12) that enable classic feed-forward architectures to build up a memory. However, LSTM delivers much better performance and can solve problems and tasks that can’t be solved by classic architectures.
LSTM networks ...
Read now
Unlock full access