A doubly differentiable function is approximated to the second degree with a Taylor expansion.
The derivative of the function must be zero for a minimum. It follows :
If the Hessian matrix is positive definite , the said zero of the derivative of is actually a minimum of and this can be approximated iteratively with the Newton method:
The problem here is that the inverse of the Hessian matrix is calculated and that it has to be positive. The quasi-Newton method is replaced by a scalar and a matrix
The derivation equation above gives transformed for and
From this we can deduce:
It is now thought that the Hessian function and are approximately the same, and concludes:
For one chooses a correction term of the form :
The equation can be rearranged so that
Thus applies
In this way the matrix can be clearly determined, but with just one correction term it is not always positive.
Davidon-Fletcher-Powell (DFP)
The matrix is approximated with the matrix and two correction terms:
properties
If is a quadratic function, with exact arithmetic the algorithm delivers the exact solution after a finite number of iterations. The following applies to all other functions
In the case of a quadratic function with parameters, the solution is ideally even achieved in steps. In practice, a little more iterations are needed, e.g. B. if the linear step size search is not carried out precisely enough or the gradients are not determined precisely enough. Usually you stop the optimization if z. B. the gradient is very small or a certain number of iterations is reached.