Given an optimization problem
In general, the question is hard to answer. We only have the following conclusion for some special objective functions and feasible sets.
Given a compact set
We now review some definitions in analysis.
For a norm function
The following figure shows the open balls of
We can define open sets and closed sets.
For closed sets, there is another different but equivalent definition.
A set
Then we define compact sets.
A set
In
A set
For optimization problems whose feasible sets are not compact, we usually cannot have simple ways to determine whether optimal solutions exist. However, for continuous function
Just like the
We first identify global minima and local minima.
Given a function
The value
Similarly, we can also define strictly global minima and strictly local minima.
Unfortunately, it is too hard to verify global minima in general. It also provides evidence why general optimization problems are difficult to solve. In this course we will study a special type of optimization problem, where local minima are also global minima.
We now give some criteria that can be used to prove local minima.
Suppose
The generalization of derivative in high dimensions is the directional derivative.
Given
In particular, if
Given
Given
In particular, if
If
If
The existence of directional derivatives cannot imply the existence of differential.
Consider the following function:
Now we give some examples and calculation rules of differentials.
Here is a simple proof of the last example:
We are ready to give the first-order optimality condition.
Suppose
An important idea is to restrict a multivariate function to a line.
Fix
Suppose
Let
In particular, if
Unfortunately, the first-order condition is a necessary condition. If
Consider function
We can compute the high-order derivatives to refute saddle points.
For a multivariate function
Given a function
We are ready to establish the second-order condition. Consider a function
Now let
Another idea is to consider the second-order Taylor series:
Suppose
In order to determine whether the Hessian of a function satisfies above condition, we introduce the definition of definite matrix.
Let
Suppose
To prove this proposition, we first introduce the eigendecomposition, which is a simplified case of SVD (singular value decomposition).
Let
For any eigenvector
We use the eigendecomposition of
Consider the following matrix
In addition, each eigenvalue
Given a matrix
Suppose
We cannot get a criterion for semidefiniteness similar to the first criterion for positive definiteness. Consider the following matrix, all of its principal minor are non-negative. Consider the following example:
Finally, we give a sufficient condition to assert a local minimum point.
Suppose
Many minimum points do not satisfy this condition. Consider the function