Skip to Content
Keras Deep Learning Cookbook
book

Keras Deep Learning Cookbook

by Rajdeep Dua, Sujit Pal, Manpreet Singh Ghotra
October 2018
Intermediate to advanced
252 pages
6h 49m
English
Packt Publishing
Content preview from Keras Deep Learning Cookbook

Optimization with AdaDelta

AdaDelta solves the problem of the decreasing learning rate in AdaGrad. In AdaGrad, the learning rate is computed as 1 divided by the sum of square roots. At each stage, we add another square root to the sum, which causes the denominator to decrease constantly. Now, instead of summing all prior square roots, it uses a sliding window that allows the sum to decrease.

AdaDelta is an extension of AdaGrad that seeks to reduce its aggressive, monotonically decreasing learning rate. Instead of accumulating all past squared gradients, AdaDelta restricts the window of accumulated past gradients to some fixed size, w.

Instead of inefficiently storing w past squared gradients, the sum of the gradients is recursively defined ...

Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Start your free trial

You might also like

Applied Deep Learning with Keras

Applied Deep Learning with Keras

Ritesh Bhagwat, Mahla Abdolahnejad, Matthew Moocarme
Advanced Deep Learning with Keras

Advanced Deep Learning with Keras

Rowel Atienza, Neeraj Verma, Valerio Maggio
The Applied TensorFlow and Keras Workshop

The Applied TensorFlow and Keras Workshop

Harveen Singh Chadha, Luis Capelo, Abhranshu Bagchi, Achint Chaudhary, Vishal Chauhan, Alexis Rutherford, Subhash Sundaravadivelu

Publisher Resources

ISBN: 9781788621755Supplemental Content