March 2018
Intermediate to advanced
272 pages
7h 53m
English
Imagine the case where we had no hidden layers and only an input and output. We talked about this architecture back in Chapter 1, The Building Blocks of Deep Learning, where we showed how it wouldn't be able to model the XOR function. Such a network architecture that wouldn't be able to model any nonlinearities in the data couldn't be modeled by the network. Each hidden layer presents an opportunity for feature engineering more and more complex interactions.
If you choose too few neurons, the outcome will likely be as follows: