This chapter will discuss a simple yet effective method for avoiding
local minima, as well as escaping from them if necessary. The
algorithm discussed is a variation of traditional simulated annealing,
optimized for the error surfaces encountered in neural network
learning. The code given here will be an implementation of that
algorithm written for general function minimization, and is primarily
for education. The specific code for learning in multiple-layer
networks can be found in the Appendix on page 429.
Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month, and much more.