Optimization
Optimization involves minimizing or maximizing a function:
If we need to maximize , we can minimize instead.
Most problems of interest, involve some constrains on what values are possible: . While optimization of arbitrary functions over arbitrary domains remains a difficult task, such problems can become somewhat simpler when and are convex.
Convexity
For simplicity, we will assume is a subset of .
A set is convex if for all and ,
Let be convex. A function is convex if for all and ,
Some important examples of convex functions include affine functions, and norms for . More generally, any norm in vector space is convex by the triangle inequality and homogeneity. We say that is strictly convex if the inequality above is replaced by a strict inequality. We can further restrict to a smaller, better behaved class of functions called strongly convex functions.
Let be convex. A function is strongly convex with parameter if is convex.
If is twice-differentiable, then -strong convexity is equivalent to the condition that for all ,
Suppose is a differentiable, convex function. If is a global minimizer of then .
Suppose further that is twice-differentiable. If is a global minimizer of , then and is positive semi-definite. Conversely, if and is positive-definite, then is a global minimizer of .
Points where are referred to as critical points.
A function is L-Lipschitz if for all ,
If is assumed to be differentiable, then for all , . If the gradient of is Lipschitz, then we say that the function is smooth.
A function is -smooth if for all , we have
Note that the -smooth condition implies that for all :