Newton’s method is one of the more successful algorithms for optimization. If it converges, it has a quadratic order of convergence. However, as pointed out before, for a general nonlinear objective function, convergence to a solution cannot be guaranteed from an arbitrary initial point *x*^{(0)} In general, if the initial point is not sufficiently close to the solution, then the algorithm may not possess the descent property for some *k*].

Recall that the idea behind Newton’s method is to locally approximate the function *f* being minimized, at every iteration, by a quadratic function. The minimizer for the quadratic approximation is used as the starting point for the next iteration. This leads to Newton’s recursive algorithm

We may try to guarantee that the algorithm has the descent property by modifying the original algorithm as follows:

where α_{k} is chosen to ensure that

For example, we may choose α_{k} = arg min_{α≥0} *f*(*x*^{(k)} − α** F**(

Start Free Trial

No credit card required