Skip to Content
Math and Architectures of Deep Learning
book

Math and Architectures of Deep Learning

by Krishnendu Chaudhury
May 2024
Intermediate to advanced content levelIntermediate to advanced
552 pages
18h 3m
English
Manning Publications
Content preview from Math and Architectures of Deep Learning

9 Loss, optimization, and regularization

This chapter covers

  • Geometrical and algebraic introductions to loss functions
  • Geometrical intuitions for softmax
  • Optimization techniques including momentum, Nesterov, AdaGrad, Adam, and SGD
  • Regularization and its relationship to Bayesian approaches
  • Overfitting while training, and dropout

By now, it should be etched in your mind that neural networks are essentially function approximators. In particular, neural network classifiers model the decision boundaries between the classes in the feature space (a space where every input feature combination is a specific point). Supervised classifiers mark sample training data inputs in this space with a—perhaps manually generated—class label (ground truth). The ...

Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Start your free trial

You might also like

Generative Deep Learning, 2nd Edition

Generative Deep Learning, 2nd Edition

David Foster
Math for Deep Learning

Math for Deep Learning

Ronald T. Kneusel

Publisher Resources

ISBN: 9781617296482Supplemental ContentPublisher SupportOtherPublisher WebsiteSupplemental ContentPurchase Link