Skip to Content
Hands-On Machine Learning for Algorithmic Trading
book

Hands-On Machine Learning for Algorithmic Trading

by Stefan Jansen
December 2018
Beginner to intermediate
684 pages
21h 9m
English
Packt Publishing
Content preview from Hands-On Machine Learning for Algorithmic Trading

Double deep Q-learning

Q-learning has been shown to overestimate action values because it purposely samples maximal estimated action values. This bias can negatively affect the learning process and the resulting policy if it does not apply uniformly and alters action preferences, as shown by Hado van Hasselt in Deep Reinforcement Learning with Double Q-learning (2015: https://arxiv.org/abs/1509.06461).

To decouple the estimation of action values from the selection of actions, Double Deep Q-Learning (DDQN) uses the weights, θ, of one network to select the best action given the next state, and the weights, θ', of another network to provide the corresponding action value estimate, that is:

One option is to randomly select one of two identical ...

Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Start your free trial

You might also like

Machine Learning for Algorithmic Trading - Second Edition

Machine Learning for Algorithmic Trading - Second Edition

Stefan Jansen

Publisher Resources

ISBN: 9781789346411Supplemental Content