Kuhn Tucker Condition
Kuhn Tucker Condition
Kuhn Tucker Condition
, 2007
You, the Kuhn-Tucker Conditions, and You
What do the Kuhn-Tucker conditions do? Simply put, they are a short cut for
writing down the first-order conditions for a constrained optimization problem
when there are also non-negativity constraints on the variables over which you
are maximizing. You could include each of the non-negativity constraints
explicitly, adding each as a constraint in the Lagrangian with an associated
Lagrange multiplier. However, in a problem with many variables, this can become
unwieldy. The Kuhn-Tucker formulation incorporates these non-negativity
constraints by imposing additional restrictions on the first-order conditions for
the problem, rather than including additional terms in the Lagrangian. But, the
important thing to remember is that the Kuhn-Tucker conditions are nothing
more than a short cut for writing down the first-order conditions for a
constrained optimization problem when there are non-negativity constraints on
the variables over which you are maximizing.
So, when you ask yourself “Do I need to use Kuhn-Tucker here?” the next question
is “Are there non-negativity constraints on the variables that I’m optimizing over?”
If the answer is “yes,” then you need to take these constraints into account when
trying to solve the problem. The Kuhn-Tucker approach is one way to do this, and
it is often the most parsimonious way to do it. On the other hand, if there are no
non-negativity constraints on the variables over which you are optimizing, then
the Kuhn-Tucker conditions are not necessary and will not help you.
1
Intuition for the Kuhn-Tucker Conditions
I’m going to focus on intuition here. What I say will be less mathematically
rigorous than a textbook presentation, but will help develop the intuition more. I
suggest using both sources to help get a full understanding of the issues.
Let’s think about a very simple problem: max f(x,y) s.t. g(x,y) ≤k and x≥0 and y≥0.
Suppose the feasible region is as in Figure 1.
y g(x,y) = k
Figure 1
The feasible region here is below the constraint and above the axes. What kinds
of solutions can there be? Well, the constraint can either bind or not, and x* = 0
or not, and y* = 0 or not. This gives us a number of cases to consider. Actually,
there are seven in total (See Figure 2).
2
y g(x,y) = k
6
7
5
x
2
4 3
Figure 2
Now, let’s take a step back and think a bit about non-negativity constraints in
general. Consider maximizing a function subject to a non-negativity constraint.
Call the function h(z). The non-negativity constraint is z≥0. There are two basic
types of solutions (See Figure 3). First, consider h1(z), with maximum at z* > 0. In
this case, the maximum is at a flat point of the function. Hence h1’(z*) = 0.
Second, consider h2(z), with maximum at z** = 0. In this case, the function could
be flat at z** = 0. But, it could also be downward sloping, as in the diagram. So,
h2’(z**) ≤ 0.
y
h1(z)
h2(z)
z** z* z
Figure 3
3
Since we don’t know whether the function h(z) looks like h1(z) or h2(z), we need a
set of first-order conditions that allows for either case. These can be written as:
either z* = 0 and h’(z*) ≤ 0, or z* > 0 and h’(z*) = 0. An even shorter way of writing
this is h’(z*) ≤ 0 and z* (h’(z*)) = 0. This is essentially what the Kunh-Tucker first-
order conditions are. They are a shorthand way of writing down these two
cases for each of the variables over which you are maximizing.
Seen in this light, the Kuhn-Tucker conditions are not all that different from the
complementary slackness conditions you already use to deal with the constraint in
the usual Lagrangian formulation. Recall, that if you have max f(x,y) s.t. g(x,y) ≤ k
(without imposing non-negativity on x and y yet), the Lagrangian is written:
and the first-order conditions are (using subscripts to denote partial derivatives):
fx(x*,y*) – λ* gx(x*,y*) = 0,
fy(x*,y*) – λ* gy(x*,y*) = 0, and
g(x*,y*) – k ≤ 0 and λ* (g(x*,y*) – k) = 0.
What’s going on with the last condition? This takes care of the idea that either
the constraint binds, or else the Lagrange multiplier is zero. Thus it is a concise
way of writing that the solution could either be in region (7) or region (1).
The Kuhn-Tucker conditions are relevant in problems where we add the non-
negativity constraints x≥0 and y≥0 to the constrained maximization problem.
They incorporate the restrictions implied by these constraints directly into the
first-order conditions. In other words, they capture the way the first-order
conditions change when the solution is in regions (2) – (6).
Intuitively, what do the Kuhn-Tucker conditions do, and where do they come from?
Let’s think about the intuition for what the additional requirements should be.
First, think about x. If x* > 0, then the derivative of the Lagrangian must be zero
(like case 1 in figure 3). (Intuitively, if the optimum is not at zero, then the non-
negativity constraint on x doesn’t matter, so the condition for the optimum on x
must be just as in the unconstrained problem.) On the other hand, if x* = 0, then
the derivative of the Lagrangian needs only to be less than or equal to zero (like
4
case 2 in figure 3). We can write these two cases with respect to x in short hand
as:
Similarly, the condition for y should allow for either y* = 0 or y* > 0. This gives us:
Since the first-order condition for the Lagrangian with respect to λ already allows
for the possibilities of the constraint binding or not, we don’t need to adjust it.
So, we are left with the Kuhn-Tucker conditions:
Let’s see how this works. What we want to do is think about the solution to the
constrained maximization problem in a particular case, and show that the Kuhn-
Tucker conditions incorporate the intuitive conditions that must hold in that case.
In doing so, we will make use of the intuition you already know for a constrained
optimization problem without non-negativity: namely that if the constraint doesn’t
bind, then the optimized value of the Lagrange multiplier (λ*) must be zero.
Consider a particular case, such as (3) above. Here we have that y = 0 and x > 0 and
g(x,y) < k. Suppose there is an optimal solution in this region. Intuitively (i.e., just
by thinking about the picture), what conditions must the optimal solution satisfy?
Since g(x*,y*) < k in this region, the Lagrange multiplier must be zero: λ* = 0. That
is, relaxing the constraint further does not affect the value of the objective
5
function. Since x* > 0, it must be that the objective function is flat with respect
to x: fx(x*,y*) = 0. Since y* = 0, it cannot be that the objective function would
increase if we increased y. Thus, it must not be that fy(x*,y*) > 0. But, this is the
same as saying that fy(x*,y*) ≤ 0. So, we have the following:
Thus, the intuitive properties of the solution in this case imply the Kuhn-Tucker
conditions. So, we have argued that if there is a solution in region (3) it satisfies
the Kuhn-Tucker conditions. You should go through each of the cases to verify
that the intuitive properties of the solution imply the Kuhn-Tucker conditions.
As a final note, recall that the Kuhn-Tucker conditions are necessary but not
sufficient. That means that any solution must satisfy them, but satisfying them
does not mean that the point actually solves the constrained maximization problem.
But, what it does mean is that if a solution exists and if there is a unique point
satisfying the Kuhn-Tucker conditions, then this is the unique solution to the
maximization problem.
6
Special Bonus Example:
y g(x,y) = k
grad f
x
(x*,y*)
Figure 4
Consider Figure 4. We have assumed that f(x,y) increases as you move up and to
the right, and we’ve drawn the indifference curves for the objective function as
dashed lines. Notice that they are steeper than the constraint at (x*,y*). This
must be the case if (x*,y*) is a maximum.
How do we formalize the idea that the indifference curve must be steeper than
the constraint at (x*,y*)? Consider the constraint: g(x,y) = k. The slope of the
constraint at a particular point is found by implicitly differentiating. Let y(x)
implicitly denote the constraint. Then we have:
g(y(x),y) ≡ k.
7
This gives us:
gx(x,y) + gy(x,y) * dy/dx = 0, or
dy/dx = - gx(x,y)/gy(x,y).
–fx(x*,y*) / fy(x*,y*).
Now, the indifference curve for f is steeper than the constraint. This means that
the absolute value of the slope of f is greater than the absolute value of the slope
of the constraint. This gives:
fx(x*,y*) / fy(x*,y*) > gx(x*,y*) / gy(x*,y*)
Now, equation (#) captures our intuition about the solution to the maximization
problem. We know (#), x* > 0, y* = 0, and g(x*,y*) = k. Do these facts imply the
Kuhn-Tucker conditions are satisfied?
First, since x* > 0, the usual first-order condition for a constrained maximization
problem should hold. So, we have:
fx(x*,y*) - λ* (gx(x*,y*)) = 0, or
8
So, to summarize, we have
It is easy to confirm that the Kuhn-Tucker first-order conditions (see the double-
lined box above) are satisfied here. So, once again we have started with
geometric intuition for what the solution must be like in a particular case (in this
case the corner solution (2)), and we’ve shown that this intuition is captured in the
Kuhn-Tucker conditions.
Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.
Alternative Proxies: