Foundations of Deep Reinforcement Learning: Theory and Practice in Python
by Laura Graesser, Wah Loon Keng
4. Deep Q-Networks (DQN)
This chapter introduces the Deep Q-Networks algorithm (DQN) proposed by Mnih et al. [88] in 2013. Like SARSA, DQN is a value-based temporal difference (TD) algorithm that approximates the Q-function. The learned Q-function is then used by an agent to select actions. DQN is only applicable to environments with discrete action spaces. However, DQN learns a different Q-function compared to SARSA—the optimal Q-function instead of the Q-function for the current policy. This small but crucial change improves the stability and speed of learning.
In Section 4.1, we first discuss why DQN learns the optimal Q-function by looking at the Bellman equation for DQN. One important implication is that this makes DQN an off-policy algorithm. ...
Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Read now
Unlock full access