Chapter 4: Convex Optimization Problems
4.1 Optimization Problem in Standard Form
An optimization problem in standard form is written as
where:
is the optimization variable. is the objective (cost) function. are inequality constraint functions. are equality constraint functions.
Optimal Value
The optimal value
: the problem is infeasible. : the problem is unbounded below.
Feasible, Optimal, and Locally Optimal Points
is feasible if it satisfies all constraints. is optimal if is feasible and . is locally optimal if is optimal for (for any ) If
is feasible and , then is called -suboptimal.
Examples (with
, : , : , , : , local optimum at
Implicit Constraints and Domain
The implicit domain constraint is
is the domain of the problem. - Explicit constraints:
, . - The problem is unconstrained if
.
Example:
is an unconstrained problem with implicit constraints
Feasibility Problem
A feasibility problem asks whether any feasible point exists:
This can be considered a special case of the general problem with
if constraints are feasible (any feasible is optimal). if infeasible.
4.2 Convex Optimization Problem
A problem is a convex optimization problem if it is of the form
Definition. An optimization problem is convex if:
is convex. are convex. are affine: .
Key Property. The feasible set of a convex optimization problem is convex: it is the intersection of
sublevel sets (of convex functions) and hyperplanes, which is convex.
4.3 Local and Global Optima
Theorem (Local = Global). For a convex optimization problem, every locally optimal point is also globally optimal.
Proof. Suppose
But local optimality requires
Uniqueness. If
is strictly convex, then a convex optimization problem has at most one global minimizer (uniqueness).
4.4 First-Order Optimality Condition
For a convex problem
Theorem (First-Order Optimality). A feasible point
is optimal if and only if
Geometric interpretation:
makes an acute ( ) angle with all feasible directions at an optimal . defines a supporting hyperplane to the feasible set at . is normal to the tangent plane of the feasible set.
Special Cases
Unconstrained optimization:
Equality-constrained minimization:
Minimization over the nonnegative orthant:
Example: Unconstrained Quadratic Minimization
Consider minimizing the quadratic function:
where
- If
, there is a unique solution: . - If
is singular and , there is no solution ( ). - If
is singular and , there are infinitely many solutions: , , where is the pseudoinverse of .
4.5 Equivalent Transformations
Two problems are (informally) equivalent if their solutions can be readily converted into each other. Common transformations that preserve equivalence:
Transformations and Change of Variables
If
This can be used to reveal "hidden convexity" of a problem.
If
Example: Log-likelihood transformation. In maximum likelihood estimation (MLE), given independent samples
applying the strictly increasing
with the key property:
Eliminating Equality Constraints
Given:
Any feasible point can be expressed as
Introducing Slack Variables
Given:
inequality constraints can be transformed via slack variables
Note. This transformation is no longer convex unless
are affine.
Epigraph Formulation
The standard form problem
is equivalent to its epigraph form:
Relaxation
Given an optimization problem
we can always take an enlarged constraint set
This is called a relaxation and its optimal value is always smaller or equal to that of the original problem (for minimization).
Example: Continuous relaxation of integer constraints.
- Feasible set becomes convex (a polytope).
- The problem reduces to a linear program (LP).
- Provides a lower bound on the integer optimum.
Example: Relaxing nonconvex sets.
Converting a Nonconvex Problem to a Convex One
Via observation:
Nonconvex problem:
Via substitution:
Nonconvex problem:
Example: Geometric program (monomial case). The original nonconvex problem
where
4.6 Linear Programming (LP)
Definition (Linear Program). An LP minimizes a linear objective subject to linear inequality and equality constraints:
An LP is a convex problem with affine objective and constraint functions. It is the most fundamental problem class in convex optimization.
Inequality and Standard Forms
Inequality form:
Standard form:
Conversion between forms:
- Replace
with , (slack variables). - Replace a free variable
with , where .
Example: Diet Problem
Find the cheapest combination of foods that satisfies certain nutritional requirements:
where
Example: Piecewise-linear Minimization
Consider minimizing the piecewise-linear function:
This can be transformed to an equivalent LP via its epigraph form:
This is an LP (in inequality form) with variables
Example: Chebyshev Center of a Polyhedron
Given a polyhedron
The problem is formulated as
is the Euclidean norm of the normal vector of the -th face. - Each constraint ensures the ball does not cross the corresponding face.
- This is a linear program (LP) in
and . - Applications: emergency facility placement, sensor/Wi-Fi placement, urban planning.
Example: Basis Pursuit
Consider an underdetermined linear system
- Exact sparsity minimization (
-norm) is NP-hard. - Relaxation: minimize the
-norm instead: s.t. . - The
-norm encourages sparsity while remaining convex.
By introducing nonnegative variables
4.7 Quadratic Programming (QP)
Definition (Quadratic Program). A QP minimizes a quadratic objective subject to linear constraints:
If
(positive semidefinite), the problem is convex. Convex QPs have globally optimal solutions and can be solved efficiently.
Example: Least Squares
Least squares problem:
Least squares with bounds:
QP formulation:
where
Example: Distance Between Two Polyhedra
Given polyhedra
This is a convex QP.
Example: Linear Program with Random Cost
Let cost be
Expanding the variance:
The QP formulation is
where
Example: LASSO (Basis Pursuit with Noise)
LASSO trades off data fidelity and sparsity.
Penalty formulation:
Constrained formulation:
Key characteristics:
- Handles noisy data (LASSO
BP when noise-free). - Sparsity controlled by
(or ). - Can be written as a QP.
4.8 Quadratically Constrained Quadratic Programming (QCQP)
Definition (QCQP). A QCQP minimizes a quadratic objective subject to quadratic inequality and linear equality constraints:
If
, the problem is convex.
QCQP generalizes QP by allowing quadratic inequality constraints, and in turn is a special case of SOCP.
4.9 Second-Order Cone Programming (SOCP)
Definition (SOCP). An SOCP is of the form:
Second-Order (Lorentz) Cone
The second-order cone (also called the Lorentz cone or ice-cream cone) is
The constraint
Why is
convex? It is the epigraph of the Euclidean norm , which is a convex function.
QP SOCP
A QP can be reformulated as an SOCP:
with
Example: Robust Linear Program (Deterministic)
In LP, there may be uncertainty in the data
Deterministic (worst-case guarantee): constraints must hold for all
Robust LP:
Equivalent SOCP:
since
Example: Robust Linear Program (Stochastic)
Assume
where
is equivalent to the SOCP (for
The terms
4.10 Semidefinite Programming (SDP)
Definition (SDP). A semidefinite program is of the form:
where
(symmetric matrices). The inequality constraint is called a linear matrix inequality (LMI).
LP SDP
An LP can be expressed as an SDP by using a diagonal matrix:
SOCP SDP
The SOC constraint
This follows from the Schur Complement Theorem: for symmetric
Thus an SOCP can be rewritten as an SDP:
Example: Eigenvalue Minimization
where
since
4.11 Hierarchy of Problem Classes
The canonical convex optimization problem classes form a nested hierarchy:
Each class is a special case of the next, with increasing modeling expressiveness and computational cost.
| Class | Objective | Constraints | Key Structure |
|---|---|---|---|
| LP | Linear | Linear inequalities | Polyhedral feasible set |
| QP | Quadratic | Linear inequalities | LP with quadratic objective |
| QCQP | Quadratic, | Quadratic inequalities | Quadratic in both objective and constraints |
| SOCP | Linear | Second-order cone constraints | |
| SDP | Linear | Linear matrix inequality | Semidefinite constraints |
Conic Program
At the top of the hierarchy is the conic program:
Definition (Conic Program).
where
is an affine map and is a convex cone.
All the previous classes are special cases of conic programs, depending on the choice of cone
| Problem | Cone |
|---|---|
| LP | |
| SOCP | |
| SDP |
4.12 Comparison: LP vs. QP in Practice
| Aspect | LP | QP |
|---|---|---|
| Objective | ||
| Constraints |
- LP
QP (QP recovers LP when ). - Often, solving an LP in theory amounts to solving a QP in practice.
Diet problem:
- In theory: cost
is deterministic LP. - In practice:
is a random vector QP.
Basis pursuit:
- In theory: samples are noise-free (
) LP. - In practice: samples are noisy (
) QP (LASSO).
Deterministic vs. stochastic robustness:
- In theory: deterministic LP
LP. - In practice: robust LP under uncertainty
SOCP.
4.13 Practical Methods for Establishing Convexity
Establishing Convexity of a Set
- Apply the definition:
, . - Show
is obtained from simple convex sets by operations that preserve convexity: - Intersection
- Affine function
- Perspective function
- Linear-fractional function
- Show
is a sublevel set of a convex function. - Show
is the epigraph of a convex function.
Establishing Convexity of a Function
- Apply the definition:
. - Use equivalent conditions (e.g.,
). - Show
is obtained from simple convex functions by operations that preserve convexity: - Nonnegative weighted sum:
, . - Composition with affine function:
. - Pointwise maximum:
. - Pointwise supremum.
- Composition with scalar functions.
- Minimization.
- Perspective.
- Nonnegative weighted sum:
4.14 Summary of Key Theorems
- The feasible set of a convex optimization problem is convex.
- Every local optimum of a convex problem is a global optimum.
- For strictly convex
, the optimal solution (if it exists) is unique. - First-order optimality condition:
is optimal for iff for all . - Problems can be made equivalent through monotone transformations, change of variables, slack variables, epigraph formulation, and relaxation.
- The hierarchy LP
QP QCQP SOCP SDP Conic Program provides progressively richer modeling frameworks, each with convex structure guaranteeing global optimality.