Solving Optimization Problems with Inequality Constraints
Similar to the solution of optimization problems with only equality constraints discussed previously, optimization problems with inequality constraints can also be solved using the Lagrange multiplier method. For a general form of the optimization problem:
minimizef(x)subjecttoh(x)=0g(x)≤0
where f:Rn→R,h:Rn→Rm,m≤n,g:Rn→Rp, we introduce the following two definitions:
Definition 1: For an inequality constraint gj(x)≤0, if at x∗ we have gj(x∗)=0, then the inequality constraint is called an active constraint at x∗; if at x∗ we have gj(x∗)<0, then the constraint is called an inactive constraint at x∗. By convention, equality constraints hi(x) are always considered active constraints.
Definition 2: Let x∗ satisfy h(x∗)=0,g(x∗)≤0, and let J(x∗) be the index set of active inequality constraints:
J(x∗)≜{j:gj(x∗)=0}
If the vectors
∇hi(x∗),∇gj(x∗),1≤i≤m,j∈J(x∗)
are linearly independent, then x∗ is called a regular point.
The following introduces the first-order necessary condition that a point must satisfy to be a local minimum, known as the KKT conditions.
KKT Conditions: Let f,h,g∈C1, and let x∗ be a regular point and a local minimum of the problem h(x)=0,g(x)≤0. Then there must exist λ∗∈Rm and μ∗∈Rp such that the following conditions hold:
When solving optimization problems with inequality constraints, one can search for points that satisfy the KKT conditions and consider these points as candidates for local minima.
Second-Order Sufficient and Necessary Conditions
Besides the first-order KKT conditions, there are second-order sufficient and necessary conditions for solving such problems.
Second-Order Necessary Condition: In the above problem, if x∗ is a local minimum and f,h,g∈C2, and assuming x∗ is a regular point, then there exist λ∗∈Rm and μ∗∈Rp such that
μ∗≥0,Df(x∗)+λ∗TDh(x∗)+μ∗TDg(x∗)=0T,μ∗Tg(x∗)=0
For all y∈T(x∗), yTL(x∗,λ∗,μ∗)y≥0 holds
Second-Order Sufficient Condition: Assume f,h,g∈C2, x∗∈Rn is a feasible point, and there exist vectors λ∗∈Rm and μ∗∈Rp such that
μ∗≥0,Df(x∗)+λ∗TDh(x∗)+μ∗TDg(x∗)=0T,μ∗Tg(x∗)=0
For all y∈T(x∗,μ∗),y=0, yTL(x∗,λ∗,μ∗)y>0 holds
Then x∗ is a strict local minimum of the optimization problem h(x)=0,g(x)≤0.