0% found this document useful (0 votes)
15 views7 pages

K Ttheory

Uploaded by

Marcelo Alonso
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
15 views7 pages

K Ttheory

Uploaded by

Marcelo Alonso
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

Mathematical Appendix I.

Kuhn-Tucker Theorems

I.1 Constrained Maximization: Necessary Conditions.

n
Function F : IR+ → IR is the objective function; functions g j : IR+
n

IR, for j = 1, . . . , k, are constraint functions. Assume that F and g j are


∂F ∂g j
differentiable, with partial derivatives ∂xi and ∂xi denoted by ∂i F and ∂i g j ,

respectively.

The constrained maximization problem (with nonnegativity constraints) is

max F (x) (1)


x

subject to g 1 (x) ≥ 0,

......,

g k (x) ≥ 0,

x1 ≥ 0, . . . , xn ≥ 0.

We write the Lagrangian as

X
k
k
1
L(λ , . . . , λ , x) = F (x) + λj g j (x),
j=1

where λj ≥ 0, for j = 1, . . . , k, are the Lagrange multipliers. We use λ to

denote the k-vector of multipliers.

1
Kuhn-Tucker conditions for x∗ ≥ 0 and λ∗ ≥ 0 are:

for all i = 1, . . . , n and j = 1, . . . , k,


X
k
∂i F (x ) +

λ∗j ∂i g j (x∗ ) ≤ 0, and if x∗i > 0, then “ = 0”, (2a)
j=1

g j (x∗ ) ≥ 0, and if λ∗j > 0, then “ = 0”. (2b)

Where do these conditions come from? Think about maximizing La-

grangian L(λ, x) with respect to x and minimizing it with respect to λ, un-

constrained, except for x ≥ 0 and λ ≥ 0. This is the saddle-point. K-T

conditions (2) are FOCs for such max-min (or saddle-point) problem.

Theorem (Kuhn-Tucker): If x∗ ≥ 0 is a solution to the constrained maxi-

mization problem, and the Constraint Qualification Condition holds, then x∗

and some λ∗ ≥ 0 satisfy K-T conditions (2).

Constraint Qualification Condition:

(i) Kuhn-Tucker original – don’t touch it.

(ii) g j concave for all j, and Slater’s condition, that is, there is some

x0 ≥ 0 with g j (x0 ) > 0 for all j.

(iii) rank condition (see Takayama 1.D.4, or Varian, ch 27),

(iv) g j linear for all j, (Arrow-Hurwicz-Uzawa, see Takayama 1.D.4)

2
I.2 Sufficiency of Kuhn-Tucker Conditions.

The most standard theorem is:

Theorem S1: Suppose that F and g 1 , . . . , g k are all concave functions. If

x∗ ≥ 0 and λ∗ ≥ 0 satisfy K-T conditions (2), then x∗ is a solution to the

constrained maximization problem.

A better theorem is due to Arrow and Enthoven (1961).

Theorem S2: Suppose that F and g 1 , . . . , g k are all quasi-concave functions

and some “mild” condition holds. If x∗ ≥ 0 and λ∗ ≥ 0 satisfy K-T conditions

(2), then x∗ is a solution to the constrained maximization problem.

The extra (“mild”) condition is not needed if F is concave (and g 1 , . . . , g k are

quasi-concave). See Takayama 1.E for three versions of the condition.

Quasi-concavity (and therefore also concavity) of functions g j implies that

the constraint set, i.e. the set of x ≥ 0 satisfying g 1 (x) ≥ 0, . . . , g k (x) ≥ 0, is

convex.

3
I.3 Constrained Minimization

The constrained minimization problem (with nonnegativity constraints) is

min F (x) (3)


x

subject to g 1 (x) ≤ 0, . . . , g k (x) ≤ 0,

x1 ≥ 0, . . . , xn ≥ 0.

The Lagrangian is
X
k
L(λ, x) = F (x) + λj g j (x).
j=1

Kuhn-Tucker conditions for x∗ ≥ 0 and λ∗ ≥ 0 are,

for all i = 1, . . . , n and j = 1, . . . , k,

X
k
∂i F (x ) +

λ∗j ∂i g j (x∗ ) ≥ 0, and if x∗i > 0, then “ = 0”, (4a)
j=1

g j (x∗ ) ≤ 0, and if λ∗j > 0, then “ = 0”. (4b)

The corresponding saddle-point problem is to minimize Lagrangian L(λ, x)

with respect to x and maximize it with respect to λ for x ≥ 0 and λ ≥ 0.

The Kuhn-Tucker Theorem holds with no change for the constrained mini-

mization problem. However, in constraint qualification conditions concavity

of functions g j , if present, has to be replaced by their convexity. This guaran-

tees convexity of the constraint set described here by inequalities g j (x) ≤ 0.

4
Theorems S1 and S2 continue to hold with concavity (quasi-concavity) of

functions F and g j replaced by their convexity (quasi-convexity, respectively).

I.4 Remarks:

• Applications of K-T theorems in microeconomics:

(i) Consumer theory: utility maximization subject to budget constraint,

and expenditure minimization.

(ii) Welfare economics: Characterization of Pareto optimal allocations as

solutions to maximization of a welfare function subject to resource con-

straints, or maximization of one agent’s utility subject to constraints on

other agents’ utilities and resource constraints.

(iii) Producer theory: cost minimization.

• There are versions of K-T theorems for maximization and minimization

with mixed constraints, i.e., when some constraints are of the equality form,

g j (x) = 0. See Sundaram [2], Section 6.4.

• K-T theorems hold for local maxima (minima) as well.

References: [1] Takayama (1995), 1.D and 1.E. [2] Sundaram (1999), Chap-

ter 6. [3] Varian, ch 27. [4] MasColell et al. Warning: Takayama [1] and

Sundaram [2] do not explicitly write nonnegativity constraints x ≥ 0. Varian

[3] writes constraints in the maximization problem as g j (x) ≤ 0.

5
I.5 Example: Consider the following constrained maximization problem:

maximize ln(x1 + 1) + ln(x2 + 1)

subject to p1 x1 + p2 x2 ≤ m

x1 ≥ 0, x2 ≥ 0,

where p1 > 0, p2 > 0 and m > 0.

In order to derive the solution (as a function of parameters p1 , p2 and m) we

write the Kuhn-Tucker first-order conditions (2) as

1
(1) − λ∗ p1 ≤ 0, and if x∗1 > 0, then “= 0”.
x∗1 +1
1
(2) − λ∗ p2 ≤ 0, and if x∗2 > 0, then “= 0”.
x∗2 +1

(3) p1 x∗1 + p2 x∗2 ≤ m, and if λ∗ > 0, then “= 0”.

with x∗ ≥ 0 and λ∗ ≥ 0.

Note that (3) holds with equality since it follows from (1) that λ∗ > 0.

We solve inequalities (1-3) by considering cases:

Case 1. x∗1 > 0, x∗2 > 0.

Then (1) and (2) hold with equalities. Solving (1), (2) and (3) we find x∗1 =
m + p2 − p1 m + p1 − p2 2
and x∗2 = and λ∗ = . For x∗1 and x∗2 to
2p1 2p2 m + p1 + p2
be strictly positive, it has to be that m + p2 > p1 and m + p1 > p2 . Thus

Case 1 applies with x∗1 and x∗2 as listed above if m + p2 > p1 and m + p1 > p2 .

6
Case 2. x∗1 > 0, x∗2 = 0.

m
(3) implies that x∗1 = . Since (1) holds with equality, we solve it for
p1
1
λ∗ = . Next we need to verify inequality (2). It states
m + p1

p2
1− ≤ 0,
m + p1

m ∗
and it holds if p2 ≥ m + p1 . Thus Case 2 applies (with x∗1 = , x = 0) if
p1 2
p2 ≥ m + p1 .

Case 3. x∗1 = 0, x∗2 > 0.

m
This case is very similar to Case 2. From (3) and (2) we obtain x∗1 = ,
p2
1
λ∗ = . Verifying inequality (1), we obtain p1 ≥ m + p2 . Thus Case 3
m + p2
m
applies (with x∗1 = 0, x∗2 = ) if p1 ≥ m + p2 .
p2

The case x∗1 = x∗2 = 0 cannot hold since it violates equation (3). This con-

cludes our solution to the K-T conditions.

Since utility function is concave and the constraint function is concave (in

fact, it is linear) K − T conditions are sufficient (Theorem S1). Hence, the

solution to K-T conditions is a constrained maximizer. Further, since the

Slater’s condition holds, every constrained maximizer has to satisfy K − T

conditions.

You might also like