Linear Constraints
Contents
10.6. Linear Constraints#
This section focuses on optimization problems where the constraints are linear (either equality or inequality constraints). We present the optimality conditions in the form of special cases of KKT conditions.
In the sequel, we present optimality conditions for the following optimization problems
Necessary optimality conditions for the minimization of a smooth cost function with linear inequality constraints.
Necessary and sufficient optimality conditions for the minimization of a convex and smooth cost function with linear inequality constraints.
Before proceeding with the optimality conditions, we present a few results which belong to a class of results called “theorems of the alternative”.
10.6.1. Farkas’ Lemma#
Farkas’ lemma is a solvability theorem for a finite system of linear inequalities. Standard versions of Farkas’ lemma consist of two different systems of linear equations and inequalities and exactly one of them is feasible.
Theorem 10.53 (Farkas’ lemma)
Let
There exists
such that .There exists
such that .
(1) and (2) are called strong alternatives as exactly one of them must be true. In contrast weak alternatives are statements in which at most one of them can be true.
Proof. We first show that both (1) and (2) cannot be true simultaneously.
Assume that both (1) and (2) are true.
Note that
since
by (1) and by (2).At the same time
since by (2)
and by (1) .Thus, we have a contradiction.
Hence (1) and (2) cannot be true simultaneously.
We now show that if (1) doesn’t hold then (2) must be true.
In other words, not (1)
Let
be the columns of .Note that the columns of
.Then, (1) can be interpreted as saying that
is a conic combination of columns of .Let
denote the cone generated by the columns of is nonempty, closed and convex cone. The closedness of is due to Theorem 9.131 since is a conic hull of a finite set of points.Since (1) doesn’t hold true, hence
.By the strict separation theorem Theorem 9.169, the set
and the point can be strictly separated by a hyperplane.Specifically, there exists
and such thatSince
, hence .Pick any
.Then,
for all since is a cone containing .Thus,
for all .Thus,
for all .Since
, hence taking the limit , we get .This holds true for every
.Choose
.Then,
.Thus,
.Also,
for all .Thus,
and as desired in (2).
Showing that if (2) doesn’t hold true, then (1) must be true is straightforward.
Assume that (2) is false.
For contradiction, assume that (1) is false.
Then (2) must be true by the previous argument.
We have a contradiction.
Hence, (1) must be true.
Theorem 10.54 (Farkas’ lemma version 2)
Let
There exists
such that .There exists
such that .
(2) is also equivalent to the following statement:
3 There exists
Proof. We first show that (2) and (3) are equivalent.
Clearly (3)
(2).Assume (2) is true.
Let
. Let . Then, .Let
.Then,
since and .Also
.Finally
.Thus,
satisfies (3).
Next, we show that (1) and (2) cannot be both true simultaneously.
For contradiction, assume that both (1) and (2) are true.
Then
since
from (1), from (2) and from (2).Also
since
from (2).We have a contradiction.
Thus, both (1) and (2) cannot be true simultaneously.
We next prove that if (2) is false then (1) must be true.
Since (2) and (3) are equivalent hence (3) is false too.
Combine the system of equations
and as whereSince (3) if false, hence there doesn’t exist
such that .This is identical to statement (1) of the original Farkas’ lemma in Theorem 10.53.
Hence statement (2) of Theorem 10.53 must hold true.
Thus, there exists
such that and .Set
where
and .Then
implies that . simplifies toThis further simplifies to
The inequality sign changes since
.Clearly,
satisfies (1) as desired. Thus, (1) is indeed true.
Theorem 10.55 (Farkas’ lemma version 3)
Let
Then the following statements are equivalent.
The implication
holds true.There exists
such that .
Proof. (2)
(2) is same as statement (1) of Theorem 10.53.
Thus, by Theorem 10.53, there is no
such that and .Hence for every
,
.Replacing
by , we see that for every ,
.
(1)
For every
, .Thus, For every
, .Thus, there doesn’t exist any
with and .Thus, by Theorem 10.53, there exists
such that .
10.6.2. Gordan’s Theorem#
Theorem 10.56 (Gordan’s theorem)
Let
. .
Proof. (1)
Assume that (1) has a solution given by
.For contradiction, assume that (2) also has a solution.
Hence, there exists a nonzero
such that and .Multiplying the equality
from left by , we getSince
and , hence every term in is negative and every term in is nonnegative.Hence
is possible only if . A contradiction.Hence (2) doesn’t have a solution.
not (1)
Assume that (1) doesn’t have a solution.
The system (1) is equivalent to the system
Define
.Let
(i.e., .Then the above system is equivalent to
The infeasibility of (1) is thus equivalent to the infeasibility of the system
where
.Since this system is infeasible, hence by Farkas' lemma, the system
must have a solution.
In other words, there exists
such thatSince
, hence .Thus, there exists
, with such that .Hence, the system (2) has a solution
.
10.6.3. Smooth Cost Function, Linear Inequality Constraints#
Theorem 10.57 (KKT necessary optimality conditions for linearly constrained problems)
Consider an optimization problem of the form
where
and
Proof. We are given that
Let
denote the constraint setBy Theorem 10.45,
must be a stationary point.Hence for every
, we haveIntroduce a change of variable
in the stationary point criterion.We must have
for every satisfying for every .Equivalently, we must have
for every satisfyingThe
-th linear inequality constraint is called active at if .The
-the constraint is called inactive if .Let the set of active constraints be denoted by
If the
-th constraint is active, then is equivalent to .If the
-th constraint is inactive, then is equivalent to a positive quantity.Thus, we must have
whenever satisfiesWe claim that the second set of inequalities are redundant.
Suppose that we pick a
such that .At
, we have for every .Then it is possible to choose a small enough
such that for every .We can choose an
for every as follows.If
, then let .If
, then letLet
.
We can now see that
But then by stationarity of
, we must have .Since
, hence we must have .
Hence, we must have
whenever satisfiesWe are ready to apply the Farkas’ lemma Theorem 10.53.
Form
by combining for every as columns.Let
.We have
.Hence the system
is infeasible.Hence there exists
such that .
Hence there exists
for every such thatBy defining
for every , we haveand
as desired.
10.6.4. Convex and Smooth Cost Function, Linear Inequality Constraints#
Theorem 10.58 (KKT necessary and sufficient optimality conditions for linearly constrained problems)
Consider an optimization problem of the form
where
and
Proof. Assume that
For the converse, assume that
Consider any feasible solution
.Define a function
By differentiating w.r.t.
, we haveBy hypothesis we have,
.Since
is convex, hence by Theorem 10.48, is a minimizer of over .By hypothesis, we have
.Hence
For any feasible point, we have
for every .Since by hypothesis
, hence for any feasible point,Hence
Hence for every feasible solution
, we have .This proves that
is a global optimal solution.
The scalars
Each multiplier
is associated with the corresponding inequality constraint .Each multiplier is nonnegative.
The conditions in (10.11) are known as complementary slackness conditions.
At the local minimizer, either the constraint is active with
or .The results can be extended to support linear equality constraints too.
10.6.5. Linear Constraints (Equality and Inequality)#
Theorem 10.59 (KKT optimality conditions for linearly constrained problems)
Consider an optimization problem of the form
where
(Necessity of KKT conditions) If
is a local minimum of this optimization problem, then there exist nonnegative scalars and real scalarssuch that
(10.13)#and
(10.14)#(Sufficiency for convex cost functions) If
is also convex over and is a feasible point of the problem (10.12) for which there exist nonnegative scalars and real scalars such that (10.13) and (10.14) are satisfied, then is an optimal solution.
The key idea for the proof is convert each linear equality constraint to two linear inequality constraints. Then we can leverage the previous results.