Chapter 12. Memory Augmented Neural Networks
So far weâve seen how effective an RNN can be at solving a complex problem like machine translation. However, weâre still far from reaching its full potential! In Chapter 9 we mentioned that itâs theoretically proven that the RNN architecture is a universal functional representer; a more precise statement of the same result is that RNNs are Turing complete. This simply means that given proper wiring and adequate parameters, an RNN can learn to solve any computable problem, which is basically any problem that can be solved by a computer algorithm or, equivalently, a Turing machine.
Neural Turing Machines
Though theoretically possible, itâs extremely difficult to achieve that kind of universality in practice. This difficulty stems from the fact that weâre looking at an immensely huge search space of possible wirings and parameter values of RNNs, a space so vastly large for gradient descent to find an appropriate solution for any arbitrary problem. However, in this chapter weâll start exploring some approaches at the edge of research that will allow us to start tapping into that potential.
Letâs think for a while about a very simple reading comprehension question like the following:
Mary travelled to the hallway. She grabbed the milk glass there. Then she travelled to the office, where she found an apple and grabbed it. How many objects is Mary carrying?
The answer is so trivial: itâs two. But what actually ...
Get Fundamentals of Deep Learning, 2nd Edition now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.