Semidefinite programming

from Wikipedia, the free encyclopedia

In semidefinite programming ( SDP , also semidefinite optimization ), optimization problems are examined whose variables are not vectors but symmetric matrices . As a secondary condition, it is required that these matrices are positive (or negative) semidefinite , from which the name of the problem results.

There are applications in the field of approximation theory , control theory , combinatorial optimization , optimal design of experiments and in technology.

Problem formulation

Given is the real vector space of the real, symmetric matrices provided with the Frobenius scalar product

.

Here the trace is a matrix.

Furthermore, let the cone of the symmetrical, positive semidefinite matrices and the generalized inequality defined by this cone , the so-called Loewner partial order, be .

Normal form

The optimization problem

with a linear semidefinite program or simply semidefinite program (SDP for short) in normal form. We are looking for a real, symmetrical matrix that is positive semidefinite, the scalar product of which assumes a certain value with the given matrices and which is maximal with respect to the Frobenius scalar product. Sometimes the equation constraints are also summarized by a linear function , which is given by

is defined. Then the inequality constraints are as .

Inequality form

Analogous to linear optimization problems, there is also the inequality form of an SDP:

where and are. Occasionally the inequality form is also written as

This corresponds to the introduction of a slip variable . This form is often chosen in order to make analogies to the linear programs clear. Here, too, a linear function is occasionally defined by

,

to simplify the notation and to make later duality statements clearer.

Without generalized inequalities

If SDP is formulated without generalized inequalities, then the conditions (normal form) and (inequality form with slip variable) are usually written out as " (or- ) is positive semidefinite".

Nonlinear semidefinite programs

Occasionally, non-linear semidefinite programs are also considered; these then either no longer have a linear objective function or have non-linear restrictions.

Classification and special cases

As convex optimization problems

Semidefinite programs are always convex optimization problems . This follows from the fact that all equation restrictions are always affine-linear and all inequality restrictions (using generalized inequalities) are always affine-linear and are therefore always K-convex functions . So the restriction set is convex. In addition, since the objective function is always linear, it is always an (abstract or generalized) convex problem, regardless of whether it is formulated as a minimization problem or a maximization problem.

As a conical program

Semidefinite programs are conical programs on the vector space of the symmetric real matrices provided with the Frobenius scalar product and using the cone of the positive semidefinite matrices. The linear subspace of the is described in the normal form by the core of the figure , i.e. by the solution set of the equation . In the inequality form with a slip variable, the subspace is described by the image of the figure .

Special case of linear programs

A special case of a semidefinite program is a linear program . To do this, all occurring matrices are replaced by diagonal matrices. This reduces the requirement that positive semidefinite should be , the Frobenius scalar product goes over to the standard scalar product and thus the equation restrictions become a linear system of equations.

example

If one wants to find a symmetric matrix for which the sum of the k largest eigenvalues ​​is as small as possible, one can formulate this as a problem of semidefinite programming. In doing so, one minimizes the variable t as the objective function, of which one demands in a secondary condition that it is greater than or equal to the sum of the k largest eigenvalues ​​of X. This constraint is very difficult to handle because there is no easy-to-calculate function that gives the eigenvalues ​​of a matrix, especially not in a sorted form. However, the constraint can be expressed equivalently by the following three conditions:

  1. .

E is the identity matrix , t and s are real variables, X and Z are matrix variables. These conditions are mathematically easier to deal with, although at first glance they look more difficult. All of them are easy to calculate because they are linear in the variables. Calculating the track is also easy. There are special procedures for checking for positive semi-definiteness for the second and third conditions, which are then used to solve the problem.

duality

Lagrange duality

Is a SDP in normal form given by

,

so the dual problem regarding the Lagrange duality can be formulated as follows. The equation constraints are formulated in order to . This gives us a Lagrange function

.

and the dual problem using the self-duality of the semidefinite cone

.

acts here as a slip variable. This is an SDP in inequality form. For the exact procedure see Lagrange duality of conical programs .

Analogous to the conical programs, one also obtains the dual problem of an SDP in inequality form

the SDP in normal form

,

Thus the SDPs are closed with regard to the Lagrange duality and the dual problem of the dual problem is always the primal problem again. In addition, the weak duality always applies , i.e. the objective function value of the dual problem is always smaller than the objective function value of the primal problem. If the Slater condition is also fulfilled (see below), then the strong duality applies , the optimal values ​​of the primal and the dual problem therefore agree.

Duality of conical programs

If SDPs are understood as abstract conical programs, then the linear subspace can be described by the linear function described above . It is then exactly the solution set of the equation . Thus, the primal conical problem can be written as SDP in normal form.

Here is . The orthogonal space required for the dual problem is then described by the operator to be adjoint , which is defined by . Thus the conical dual problem is:

.

The dual problem is then an SDP in the form of an inequality with a slip variable. It then always applies to all permissible ones

If the Slater condition is fulfilled (see below) and the primal problem has a finite optimal value , then the dual problem has an optimal solution and it holds

.

Slater condition

The Slater condition is a requirement of the primal problem that guarantees that strong duality holds. It requires that the problem has a point which satisfies the equation constraints and strictly satisfies all inequality constraints, that is, or for at least one point in all inequality constraints of the problem simultaneously. But since if is, which in turn is equivalent to being a positively definite matrix, the Slater condition for SDPs is already fulfilled if there is a positively definite matrix that fulfills the secondary equation conditions.

literature

  • Florian Jarre, Josef Stoer: Optimization. Springer, Berlin 2004, ISBN 3-540-43575-1 .
  • Johannes Jahn: Introduction to the Theory of Nonlinear Optimization. 3. Edition. Springer, Berlin 2007, ISBN 978-3-540-49378-5 .

Individual evidence

  1. Florian Jarre, Josef Stoer: Optimization. Springer, Berlin 2004, p. 419.

Web links