Skip to Content
Java Deep Learning Cookbook
book

Java Deep Learning Cookbook

by Rahul Raj
November 2019
Intermediate to advanced
304 pages
8h 40m
English
Packt Publishing
Content preview from Java Deep Learning Cookbook

Determining the right activation function

The purpose of an activation function is to introduce non-linearity into a neural network. Non-linearity helps a neural network to learn more complex patterns. We will discuss some important activation functions, and their respective DL4J implementations.

The following are the activation functions that we will consider:

  • Tanh
  • Sigmoid
  • ReLU (short for Rectified Linear Unit)
  • Leaky ReLU
  • Softmax

In this recipe, we will walk through the key steps to decide the right activation functions for a neural network.

Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Start your free trial

You might also like

Java Deep Learning Projects

Java Deep Learning Projects

Md. Rezaul Karim
Java: Data Science Made Easy

Java: Data Science Made Easy

Richard M. Reese, Jennifer L. Reese, Alexey Grigorev
Java 9 High Performance

Java 9 High Performance

Mayur Ramgir, Nick Samoylov
Introduction to Deep Learning Using PyTorch

Introduction to Deep Learning Using PyTorch

Goku Mohandas, Alfredo Canziani

Publisher Resources

ISBN: 9781788995207Supplemental Content