Trust region procedure

from Wikipedia, the free encyclopedia

The trust region methods are a class of robust and efficient globalization strategies for the numerical calculation of a local minimum of a possibly non-convex, one time continuously differentiable function. The trust region methods are closely related to the Levenberg-Marquardt algorithm , but have significant differences in the quadratic sub-problems to be solved and are deterministic in the choice of the step size restriction. Under certain circumstances it can also be shown that line search methods are special variants of trust region methods.

Overview

Trust region methods are used to solve nonlinear programming problems (NLP) of the type

to solve.

Here it is assumed that once is continuously differentiable on . But that doesn't mean that it is a convex function . Since you can also solve the following non-smooth nonlinear programming problem with the slightest changes to the algorithm

solve, we will consider the trust region method for this class of problems. Here is with , as well as a Hilbert space.

A trust region procedure is an iterative procedure which consists of individual, so-called trust region steps. Such a trust region step takes place in two logical units: First, a correction is calculated. The way in which such a correction is calculated can in principle be freely selected, as long as the correction fulfills a so-called "Sufficient Decrease Condition". For performance reasons, however, the correction is often calculated by solving a quadratic minimization problem under constraints. The corrections calculated with this Sequential Quadratic Programming (SQP) variant are under certain circumstances the corrections that are calculated in a (quasi-) Newton step. Since this correction can be arbitrarily bad in terms of the task of the minimization problem, the usefulness of the correction is measured and the secondary conditions of the quadratic problem are changed on the basis of this.

A trust region step in detail

Calculate the correction

We assume that a feasible iterate is given. How to calculate a correction by solving the following Quadratic Programming Problem (QP):

with the quadratic function

Here refers to the trust region radius and a symmetric approximation to the possibly existing Hessian matrix of . In this case the function is a second order Taylor approximation an .

We assume briefly that the matrix is symmetrically positive (semi-) definite and that there are no constraints in the above QP problem. So the necessary and also sufficient conditions for a minimum are straight

which is a quasi-Newton step .

Comment on the solution of the quadratic problem

This minimization problem can generally be solved approximately. This means that the solution just has to be a better solution to the QP problem than the so-called Cauchy point . More precisely, this means that the following inequality must be satisfied

where is a Coleman-Li scaling matrix that stores the distance to the obstacle on the main diagonal .

Without further assumptions one must use a formulation with a penalty function for the solution of the quadratic minimization problem in order to be included in the solution. However, for a finite dimensional space , can be replaced by and a truncated conjugate gradient method (truncated CG) or a nonlinear Gauss-Seidel method can be used for the solution.

As mentioned, a correction can be as bad as you want, so a scalar, the so-called contraction rate, is calculated to determine the quality of a correction

The value measures the deviation of the quadratic function by the predicted reduction of (predicted reduction) of the real (actual reduction). It is therefore often found in the literature

Comment on : In fact, the rate of contraction measures the linearity of the gradient of . If is a quadratic function, the rate of contraction will always be 1, provided . In practice, you should therefore also test whether this applies.

Update step

Finally, it is determined on the basis of whether the correction is accepted and how the next trust region radius is chosen.

If so, and is chosen. Otherwise is and

Here and are to be chosen a priori. In the literature, other constants are often used to fine-tune the choice of the new trust region radius, but the method manages with these constants.

Convergence properties

Under certain but very weak assumptions one can show that the iterates calculated in this way converge to solutions of the necessary first-order conditions.

Basically you proceed as follows:

  • the solution of the QP problem always satisfies the sufficient decrease condition (if not one chooses the Cauchy point). That means: If corrections are successful, i.e. if the following inequality applies

So you can limit the actual descent in by the norm of the gradient and the trust region radius downwards.

  • If one now assumes that the norm of the gradient is bounded by below, one obtains the following: Assuming there are infinitely many successful steps, then converges to or and one solves the problem (at least its necessary first-order conditions). Otherwise, the trust region radius must converge to 0 due to the above inequality. In that case, however, the quadratic function would provide an increasingly better approximation and the decrease ratio would approach 1. But that would contradict the assumption that there are not an infinite number of successful steps.

Differences to other procedures

  • The Levenberg-Marquardt method performs a nondeterministic update of the step size restriction.
  • The trust region method is Newton-like; H. the corrections are calculated using a second-order Taylor expansion; the Levenberg-Marquardt method solves a quadratic auxiliary problem based on a first-order Taylor expansion.
  • In contrast to the line search method, in the trust region method, the step size restriction is selected before a correction is calculated. Due to the additional constraints in the minimization problem , a different and better solution is calculated than the same step size restriction in the line search algorithm would provide.

Enhancements to trust region procedures

  • To solve nonlinear programming problems with even more general constraints of the kind
wherein can be so-called filter-Trust Region method use.
  • There are extensions of trust region procedures that also calculate convergence to critical points of the second order, i.e. points for which applies
and .

Methods based on trust region procedures

literature

  • Thomas F. Coleman, Yuying Li: On the convergence of interior-reflective Newton methods for nonlinear minimization subject to bounds. In: Math. Programming. 67 (2, Ser. A) 1994, ISSN  0025-5610 , pp. 189-224.
  • AR Conn, NIM Gould, Ph. L. Toint: Trust-region methods . Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 2000.

Individual evidence

  1. ^ A b Thomas F. Coleman, Yuying Li: On the convergence of interior-reflective Newton methods for nonlinear minimization subject to bounds.
  2. AR Conn, NIM Gould, Ph. L. Toint: Trust-region methods.