0% found this document useful (0 votes)
21 views797 pages

Slides

The document outlines a course on economic decision-making, emphasizing the importance of individual choices under certainty and uncertainty. It discusses foundational economic theories from Adam Smith and Jeremy Bentham, introduces preference relations, and explores decision-making models. The course aims to develop a comprehensive understanding of choice theory, general equilibrium, and the limitations of welfare theorems.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views797 pages

Slides

The document outlines a course on economic decision-making, emphasizing the importance of individual choices under certainty and uncertainty. It discusses foundational economic theories from Adam Smith and Jeremy Bentham, introduces preference relations, and explores decision-making models. The course aims to develop a comprehensive understanding of choice theory, general equilibrium, and the limitations of welfare theorems.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 797

M100

Matthew Elliott
Cambridge and Caltech

October, 2016
Some Logistics

These slides borrow heavily from teaching notes by Clayton


Featherstone, Jonathan Levin, Paul Milgrom, Antonio Rangel,
Ilya Segal and especially Luke Stein.

My office hour is after class on Thursdays (Office 42).

I’ll be available to chat after class most of the time.

You are lucky you have an outstanding TA, River. Please use
him!

Handouts are intended to allow you to concentrate on the


material in lectures.
Who said this?

“Nobody ever saw a dog make a fair and deliberate exchange of


one bone for another with another dog. Nobody ever saw one
animal by its gestures and natural cries signify to another, this is
mine, that yours; I am willing to give this for that....But man has
almost constant occasion for the help of his brethren, and it is in
vain for him to expect it from their benevolence only. He will be
more likely to prevail if he can interest their self-love in his favour,
and show them that it is for their own advantage to do for him
what he requires of them. Whoever offers to another a bargain of
any kind, proposes to do this. Give me that which I want, and you
shall have this which you want, is the meaning of every such offer;
and it is in this manner that we obtain from one another the far
greater part of those good offices which we stand in need of.”
Or, more famously

“It is not from the benevolence of the butcher, the brewer, or the
baker that we expect our dinner, but from their regard to their own
interest.”
Or, more famously

“It is not from the benevolence of the butcher, the brewer, or the
baker that we expect our dinner, but from their regard to their own
interest.”

Adam Smith, The Wealth of Nations (1776).


Who said this?

“The greatest happiness of the greatest number is the foundation


of morals and legislation.”
Who said this?

“The greatest happiness of the greatest number is the foundation


of morals and legislation.”
“It is vain to talk of the interest of the community, without
understanding what is the interest of the individual.”
Who said this?

“The greatest happiness of the greatest number is the foundation


of morals and legislation.”
“It is vain to talk of the interest of the community, without
understanding what is the interest of the individual.”

Jeremy Bentham (1748-1832).


Economic thinking in early 19th Century

Want to refine Adam Smith’s ideas about an economic system


based on self-interest.

Perhaps the project could be advanced by an index of


self-interest?

How beneficial are different outcomes to any individual?


Utilitarian philosophers in early 19th Century

Wanted an objective criterion for a science of government.

But requires a way of measuring how beneficial different


policies are to different people.
Utilitarian philosophers in early 19th Century

Wanted an objective criterion for a science of government.

But requires a way of measuring how beneficial different


policies are to different people.

Utility-maximization theory of decision making is born.


Individual Decision-Making Under Certainty

Our study begins with individual decision-making under certainty


Individual Decision-Making Under Certainty

Our study begins with individual decision-making under certainty

We start with a very general problem that only includes:


Feasible set (what can you choose from)
Choice correspondence (what you choose)
Individual Decision-Making Under Certainty

Our study begins with individual decision-making under certainty

We start with a very general problem that only includes:


Feasible set (what can you choose from)
Choice correspondence (what you choose)

A fairly innocent assumption will then allow us convert this


problem into an optimization problem (more later).
Individual Decision-Making Under Certainty

Our study begins with individual decision-making under certainty

We start with a very general problem that only includes:


Feasible set (what can you choose from)
Choice correspondence (what you choose)

A fairly innocent assumption will then allow us convert this


problem into an optimization problem (more later).

We’ll generalize to decision making under uncertainty later.


What will the model deliver?

Useful:
Can often recover preferences from choices.
Aggregating into a theory of government it is aligned with
democratic values.
But. . . interpersonal comparisons prove difficult
Accurate (somewhat): many predictions are empirically
verified.
Broad
Consumption and production
Lots of other things
Compact
Extremely compact formulation
Ignores an array of other important “behavioral” factors
How we will use it: A road map

(i) Choice under certainty.


(ii) Choice under uncertainty.
Our theory of decision making can be applied to study:
(i) Consumers’ purchasing decisions.
(ii) Producers’ supply decisions.
(iii) General Equilibrium
Ultimately, this will allow us to model markets and the
economy.
We’ll be able to rigorously flesh out Adam Smith’s ideas
building a model in which the “invisible hand” delivers.
(iv) Limitations
Our general equilibrium analysis will rely on (strong)
assumptions. We’ll start to relax these.
Outline

1 Choice Theory
Preferences
Utility functions
Properties of preferences
Behavioral critiques
Comparative statics
An application: Consumer Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Preference relations

Definition (weak preference relation)


 is a binary relation on a set of possible choices X such that
x  y iff “x is at least as good as y.”

How might we use this relation to define strict preferences and


indifference?
Preference relations

Definition (weak preference relation)


 is a binary relation on a set of possible choices X such that
x  y iff “x is at least as good as y.”

How might we use this relation to define strict preferences and


indifference?
Definition (strict preference relation)
 is a binary relation on X such that x  y (“x is strictly
preferred to y”) iff x  y but y 6 x.

Definition (indifference)
∼ is a binary relation on X such that x ∼ y (“the agent is
indifferent between x and y”) iff x  y and y  x.
Properties of preference relations

What properties might we expect  to satisfy?


Properties of preference relations

What properties might we expect  to satisfy?


Definition (completeness)
 on X is complete iff ∀x, y ∈ X, either x  y or y  x.

Completeness implies that x  x


Definition (transitivity)
 on X is transitive iff whenever x  y and y  z, we have x  z.

Rules out preference cycles except in the case of indifference


Properties of preference relations

What properties might we expect  to satisfy?


Definition (completeness)
 on X is complete iff ∀x, y ∈ X, either x  y or y  x.

Completeness implies that x  x


Definition (transitivity)
 on X is transitive iff whenever x  y and y  z, we have x  z.

Rules out preference cycles except in the case of indifference


Definition (rationality)
 on X is rational iff it is both complete and transitive.
Example from Kahneman and Tversky (84)

Suppose you are about to buy a stereo for $125 and a


calculator for $15.

There is a $5 discount on the calculator at a store 10 minutes


away. Do you make the trip? (Yes)

There is a $5 discount on the stereo at a store 10 minutes


away. Do you make the trip? (No)
Example from Kahneman and Tversky (84)

You find out both items are out of stock and need to go to
the other branch 10 minutes away. You get a $5 dollar
discount as compensation. Which item would you rather have
the discount on? (Indifferent).
Example from Kahneman and Tversky (84)

You find out both items are out of stock and need to go to
the other branch 10 minutes away. You get a $5 dollar
discount as compensation. Which item would you rather have
the discount on? (Indifferent).

Which property of rational choice has been violated?


Example from Kahneman and Tversky (84)

Suppose you are about to buy a stereo for $125 and a


calculator for $15.

There is a $5 discount on the calculator at a store 10 minutes


away. Do people make the trip? (Yes)

There is a $5 discount on the stereo at a store 10 minutes


away. Do people make the trip? (No)
Example from Kahneman and Tversky (84)

Suppose you are about to buy a stereo for $125 and a


calculator for $15.

There is a $5 discount on the calculator at a store 10 minutes


away. Do people make the trip? (Yes)

There is a $5 discount on the stereo at a store 10 minutes


away. Do people make the trip? (No)
Consider the following three options:
(i) staying at the store, no discount.
(ii) Traveling 10 minutes for the calculator discount.
(iii) Traveling 10 minutes for the stereo discount.

From the above choices what can we infer: (ii)  (i) and
(i)  (iii).
Example from Kahneman and Tversky (84)

You find out both items are out of stock and need to go to
the other branch 10 minutes away. You get a $5 dollar
discount as compensation. Which item would you rather have
the discount on? (Indifferent).
Recall the following three options:
(i) staying at the store, no discount.
(ii) Traveling 10 minutes for the calculator discount.
(iii) Traveling 10 minutes for the stereo discount.
From the above choices what can we infer: (ii) ∼ (iii).
Example from Kahneman and Tversky (84)

You find out both items are out of stock and need to go to
the other branch 10 minutes away. You get a $5 dollar
discount as compensation. Which item would you rather have
the discount on? (Indifferent).
Recall the following three options:
(i) staying at the store, no discount.
(ii) Traveling 10 minutes for the calculator discount.
(iii) Traveling 10 minutes for the stereo discount.
From the above choices what can we infer: (ii) ∼ (iii).
From before (ii)  (i) and (i)  (iii).
Example from Kahneman and Tversky (84)

You find out both items are out of stock and need to go to
the other branch 10 minutes away. You get a $5 dollar
discount as compensation. Which item would you rather have
the discount on? (Indifferent).
Recall the following three options:
(i) staying at the store, no discount.
(ii) Traveling 10 minutes for the calculator discount.
(iii) Traveling 10 minutes for the stereo discount.
From the above choices what can we infer: (ii) ∼ (iii).
From before (ii)  (i) and (i)  (iii).
(ii)  (i)  (iii)  (ii), so by transitivity (ii) ∼ (i) ∼ (iii).
Example from Kahneman and Tversky (84)

You find out both items are out of stock and need to go to
the other branch 10 minutes away. You get a $5 dollar
discount as compensation. Which item would you rather have
the discount on? (Indifferent).
Recall the following three options:
(i) staying at the store, no discount.
(ii) Traveling 10 minutes for the calculator discount.
(iii) Traveling 10 minutes for the stereo discount.
From the above choices what can we infer: (ii) ∼ (iii).
From before (ii)  (i) and (i)  (iii).
(ii)  (i)  (iii)  (ii), so by transitivity (ii) ∼ (i) ∼ (iii).
Contradicts (ii)  (i) (and (i)  (iii)), transitivity violated.
Summary of preference notation

yx y 6 x
xy x∼y xy
x 6 y yx Ruled out by completeness
Summary of preference notation

yx y 6 x
xy x∼y xy
x 6 y yx Ruled out by completeness

Can think of (complete) preferences as inducing a function

p : X × X → {, ∼, ≺}
Other properties of rational preference relations

Assume  is rational. Then for all x, y, z ∈ X:


Weak preference is reflexive: x  x
Indifference is
Reflexive: x ∼ x
Transitive: (x ∼ y) ∧ (y ∼ z) =⇒ x ∼ z
Symmetric: x ∼ y ⇐⇒ y ∼ x
Strict preference is
Irreflexive: x 6 x
Transitive: (x  y) ∧ (y  z) =⇒ x  z
(x  y) ∧ (y  z) =⇒ x  z, and
(x  y) ∧ (y  z) =⇒ x  z
Approaches to modelling individual decision-making

1 Conventional approach
Start from preferences, ask what choices are compatible
Approaches to modelling individual decision-making

1 Conventional approach
Start from preferences, ask what choices are compatible

2 Revealed-preference approach
Start from observed choices, ask what preferences are
compatible
What can we learn about underlying preferences from
observing choices?
Can we test rational choice theory? How?
Are all decisions consistent with some rational preferences?
Choice rules

Definition (Choice rule)


Given preferences  over X, and choice set B ⊆ X, the choice
rule is a correspondence giving the set of all “best” elements in B:

C(B, ) ≡ {x ∈ B : x  y for all y ∈ B}.


Choice rules

Definition (Choice rule)


Given preferences  over X, and choice set B ⊆ X, the choice
rule is a correspondence giving the set of all “best” elements in B:

C(B, ) ≡ {x ∈ B : x  y for all y ∈ B}.

Theorem
Suppose  is complete and transitive and B finite and non-empty.
Then C(B, ) 6= ∅.
Discussion of theorem

Conclusion of the Theorem says that the choice rule never


selects nothing.
Does this make sense?
What if I don’t like any of the things I am choosing between?
Why can’t my most preferred choice be choosing nothing?
Discussion of theorem

Conclusion of the Theorem says that the choice rule never


selects nothing.
Does this make sense?
What if I don’t like any of the things I am choosing between?
Why can’t my most preferred choice be choosing nothing?
Choice rule is about choosing among items in B only.
The action of choosing nothing could be included in B.
Theorem is a “consistency check” for the choice rule.
Outline of the proof for non-emptiness of choice
correspondence

Proof is by mathematical induction. This is a very useful


technique. Here is an outline of the argument.
With one item B = x, by completeness x  x, so it would be
selected and x ∈ C(A, ).
Outline of the proof for non-emptiness of choice
correspondence

Proof is by mathematical induction. This is a very useful


technique. Here is an outline of the argument.
With one item B = x, by completeness x  x, so it would be
selected and x ∈ C(A, ).
Suppose now there are n items and the choice set is not empty.
Outline of the proof for non-emptiness of choice
correspondence

Proof is by mathematical induction. This is a very useful


technique. Here is an outline of the argument.
With one item B = x, by completeness x  x, so it would be
selected and x ∈ C(A, ).
Suppose now there are n items and the choice set is not empty.
We will show that with n + 1 items the choice set is not empty.
Outline of the proof for non-emptiness of choice
correspondence

Proof is by mathematical induction. This is a very useful


technique. Here is an outline of the argument.
With one item B = x, by completeness x  x, so it would be
selected and x ∈ C(A, ).
Suppose now there are n items and the choice set is not empty.
We will show that with n + 1 items the choice set is not empty.
Without loss of generality, suppose x ∈ C(B, ). Let
A = B ∪ {y}. By completeness either x  y, in which case
x ∈ C(A, ), or y  x. By transitivity y must then be preferred to
all elements of A, so y ∈ C(A, ).
Outline of the proof for non-emptiness of choice
correspondence

Proof is by mathematical induction. This is a very useful


technique. Here is an outline of the argument.
With one item B = x, by completeness x  x, so it would be
selected and x ∈ C(A, ).
Suppose now there are n items and the choice set is not empty.
We will show that with n + 1 items the choice set is not empty.
Without loss of generality, suppose x ∈ C(B, ). Let
A = B ∪ {y}. By completeness either x  y, in which case
x ∈ C(A, ), or y  x. By transitivity y must then be preferred to
all elements of A, so y ∈ C(A, ).
Letting n = 1, we have proved it for n = 2. Letting n = 2 . . . , and
so on.
Revealed preference

Before: used known preference relation  to generate choice


rule C(·, )
Now: suppose agent reveals her preferences through her
choices, which we observe; can we deduce a rational
preference relation that could have generated them?
Revealed preference

Before: used known preference relation  to generate choice


rule C(·, )
Now: suppose agent reveals her preferences through her
choices, which we observe; can we deduce a rational
preference relation that could have generated them?

Definition (revealed preference choice rule)


Any CR : 2X → 2X (where 2X means the set of subsets of X)
such that for all A ⊆ X, we have CR (A) ⊆ A.

If CR (·) could be generated by a rational preference relation (i.e.,


there exists some complete, transitive  such that
CR (A) = C(A, ) for all A), we say it is rationalizable
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c}
{a} {a}
{a, b} {c}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a}
{a, b} {c}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
∅ {c}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
∅ {c} X (Not possible)
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
∅ {c} X (Not possible)
{b} {a}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
∅ {c} X (Not possible)
{b} {a} X (Not possible)
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
∅ {c} X (Not possible)
{b} {a} X (Not possible)
{a} {a, b}
Examples of revealed preference choice rules

Suppose we know CR (·) for


A ≡ {a, b}
B ≡ {a, b, c}

CR ({a, b}) CR ({a, b, c}) Possibly rationalizable?


{a} {c} X (c  a  b)
{a} {a} X (a  b, a  c, b?c)
{a, b} {c} X (c  a ∼ b)
{c} {c} X (c 6∈ {a, b})
∅ {c} X (Not possible)
{b} {a} X (Not possible)
{a} {a, b} X (Not possible)
A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
b ∈ A (b could have been chosen)
A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
b ∈ A (b could have been chosen)

We can infer that a  b


A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
b ∈ A (b could have been chosen)

We can infer that a  b

Now consider some B ⊆ X such that


a∈B
b ∈ CR (B) (b was chosen)
So: b  z for all z ∈ B
A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
b ∈ A (b could have been chosen)

We can infer that a  b

Now consider some B ⊆ X such that


a∈B
b ∈ CR (B) (b was chosen)
So: b  z for all z ∈ B

We can infer that b  a


A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
b ∈ A (b could have been chosen)

We can infer that a  b

Now consider some B ⊆ X such that


a∈B
b ∈ CR (B) (b was chosen)
So: b  z for all z ∈ B

We can infer that b  a

Thus a ∼ b
A necessary condition for rationalizability

Suppose that CR (·) is rationalizable (in particular, it is generated


by ), and we observe CR (A) for some A ⊆ X such that
a ∈ CR (A) (a was chosen).
So: a  z for all z ∈ A
b ∈ A (b could have been chosen)

We can infer that a  b

Now consider some B ⊆ X such that


a∈B
b ∈ CR (B) (b was chosen)
So: b  z for all z ∈ B

We can infer that b  a

Thus a ∼ b, hence a ∈ CR (B) and b ∈ CR (A) by transitivity


Houthaker’s Axiom of Revealed Preferences

A rationalizable choice rule CR (·) must therefore satisfy “HARP”:


Definition (Houthaker’s Axiom of Revealed Preferences)
Revealed preferences CR : 2X → 2X satisfies HARP iff ∀a, b ∈ X
and ∀A, B ⊆ X such that
{a, b} ⊆ A and a ∈ CR (A); and
{a, b} ⊆ B and b ∈ CR (B),
we have that a ∈ CR (B) (and b ∈ CR (A)).
Houthaker’s Axiom of Revealed Preferences

A rationalizable choice rule CR (·) must therefore satisfy “HARP”:


Definition (Houthaker’s Axiom of Revealed Preferences)
Revealed preferences CR : 2X → 2X satisfies HARP iff ∀a, b ∈ X
and ∀A, B ⊆ X such that
{a, b} ⊆ A and a ∈ CR (A); and
{a, b} ⊆ B and b ∈ CR (B),
we have that a ∈ CR (B) (and b ∈ CR (A)).

It turns out HARP is not only necessary, but sufficient for


rationalizability
Illustrating HARP

A violation of HARP:
Example of HARP

Suppose
1 Revealed preferences CR (·) satisfy HARP, and that
2 CR (·) is nonempty-valued (except for CR (∅))

If CR ({a, b}) = {b}, what can we conclude about


CR ({a, b, c})?

If CR ({a, b, c}) = {b}, what can we conclude about


CR ({a, b})?
Example of HARP

Suppose
1 Revealed preferences CR (·) satisfy HARP, and that
2 CR (·) is nonempty-valued (except for CR (∅))

If CR ({a, b}) = {b}, what can we conclude about


CR ({a, b, c})?

CR ({a, b, c}) ∈ {{b}, {c}, {b, c}}

If CR ({a, b, c}) = {b}, what can we conclude about


CR ({a, b})?
Example of HARP

Suppose
1 Revealed preferences CR (·) satisfy HARP, and that
2 CR (·) is nonempty-valued (except for CR (∅))

If CR ({a, b}) = {b}, what can we conclude about


CR ({a, b, c})?

CR ({a, b, c}) ∈ {{b}, {c}, {b, c}}

If CR ({a, b, c}) = {b}, what can we conclude about


CR ({a, b})?
CR ({a, b}) = {b}
HARP is necessary and sufficient for rationalizability

Theorem
Suppose revealed preference choice rule CR : 2X → 2X is
nonempty-valued (except for CR (∅)). Then CR (·) satisfies HARP
iff there exists a rational preference relation  such that
CR (·) = C(·, ).
Proof Logic

Already argued that rationality implies HARP.


To see HARP implies rationality we construct a preference
relation.
For any x and y, let x c y iff there exists some A ⊆ X such
that y ∈ A and x ∈ CR (A).

Can consider all pairs of x, y, so constructed relation is


complete.
HARP then implies that transitivity holds, so relation is
rational.
Revealed preference and limited data

Our discussion relies on all preferences being observed

All elements of CR (A)

CR (A) for every A ⊆ X


Revealed preference and limited data

Our discussion relies on all preferences being observed. . . real data


is typically more limited
All elements of CR (A). . . we may only see one element of A
eR (A) ∈ CR (A)
i.e., C
CR (A) for every A ⊆ X
Revealed preference and limited data

Our discussion relies on all preferences being observed. . . real data


is typically more limited
All elements of CR (A). . . we may only see one element of A
eR (A) ∈ CR (A)
i.e., C
CR (A) for every A ⊆ X. . . we may only observe choices for
certain choice sets
bR (A) : B → 2X for B ⊂ 2X with C
i.e., C bR (A) = CR (A)
Revealed preference and limited data

Our discussion relies on all preferences being observed. . . real data


is typically more limited
All elements of CR (A). . . we may only see one element of A
eR (A) ∈ CR (A)
i.e., C
CR (A) for every A ⊆ X. . . we may only observe choices for
certain choice sets
bR (A) : B → 2X for B ⊂ 2X with C
i.e., C bR (A) = CR (A)

Other “axioms of revealed preference” hold in these environments


Weak Axiom of Revealed Preference (WARP)
Generalized Axiom of Revealed Preference (GARP)—necessary
and sufficient condition for rationalizability
Weak Axiom of Revealed Preferences

Definition (Weak Axiom of Revealed Preferences)


bR : B → 2X defined only for choice sets
Revealed preferences C
X
B ⊆ 2 satisfies WARP iff ∀a, b ∈ X and ∀A, B ∈ B such that
{a, b} ⊆ A and a ∈ CbR (A); and
{a, b} ⊆ B and b ∈ C
bR (B),
we have that a ∈ C
bR (B) (and b ∈ C
bR (A)).
Weak Axiom of Revealed Preferences

Definition (Weak Axiom of Revealed Preferences)


bR : B → 2X defined only for choice sets
Revealed preferences C
X
B ⊆ 2 satisfies WARP iff ∀a, b ∈ X and ∀A, B ∈ B such that
{a, b} ⊆ A and a ∈ CbR (A); and
{a, b} ⊆ B and b ∈ C
bR (B),
we have that a ∈ C
bR (B) (and b ∈ C
bR (A)).

HARP is WARP with all possible choice sets (i.e,. B = 2X )

WARP is necessary but not sufficient for rationalizability


Weak Axiom of Revealed Preferences cont.

Example
bR : B → 2{a,b,c} defined for choice sets
Consider C
B ≡ {{a, b}, {b, c}, {c, a}} ⊆ 2{a,b,c} with:
CbR ({a, b}) = {a},
bR ({b, c}) = {b}, and
C
bR ({c, a}) = {c}.
C

C
bR (·) satisfies WARP, but is not rationalizable.

Think of CbR (·) as a restriction of some CR : 2{a,b,c} → 2{a,b,c} ;


there is no CR ({a, b, c}) consistent with HARP.
Outline

1 Choice Theory
Preferences
Utility functions
Properties of preferences
Behavioral critiques
Comparative statics
An application: Consumer Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


From abstract preferences to maximization

Our model of choice so far is entirely abstract


Utility assigns a numerical ranking to each possible choice
By assigning a utility to each element of X, we turn the
choice problem into an optimization problem

Definition (utility function)


Utility function u : X → R represents  on X iff for all x, y ∈ X,

x  y ⇐⇒ u(x) ≥ u(y).

Then the choice rule is

C(B, ) ≡ {x ∈ B : x  y for all y ∈ B} = argmax u(x)


x∈B
Utility representation implies rationality

Theorem
If utility function u : X → R represents  on X, then  is rational.

Can show  is complete. How?

Can show  is transitive. How?


Ordinality of utility and interpersonal comparisons

Note that  is represented by any function satisfying

x  y ⇐⇒ u(x) ≥ u(y)

for all x, y ∈ X

Thus any increasing monotone transformation of u(·) also


represents 
The property of representing  is ordinal

There is no such thing as a “util”


Failure of interpersonal comparisons

Interpersonal comparisons are impossible using this theory

1 Disappointing to original utilitarian agenda

2 Rawls (following Kant, following. . . ) attempts to solve this by


asking us to consider only a single chooser
Can we find a utility function representing ?

Theorem
Any complete and transitive preference relation  on a finite set X
can be represented by some utility function u : X → {1, . . . , n}
where n ≡ |X|.

Intuitive argument—a constructive algorithm:


1 Assign the “top” elements of X utility n = |X|
2 Discard them; we are left with a set X 0
3 If X 0 = ∅, we are done; otherwise return to step 1 with the set
X0
What if |X| = ∞?

If X is infinite, our proof doesn’t go through, but we still may be


able to represent  by a utility function

Example
Preferences over R+ with x1  x2 iff x1 ≥ x2 .
 can be represented by u(x) = x. (It can also be represented by
other utility functions.)

However, if X is infinite we can’t necessarily represent  by a


utility function
Example (lexicographic preferences)
Preferences over [0, 1]2 ⊆ R2 with (x1 , y1 )  (x2 , y2 ) iff
x1 > x2 , or
x1 = x2 and y1 ≥ y2 .
Continuous Preferences

Definition (continuous preference relation)


A preference relation  on X is continuous iff for any sequence
{(xn , yn )}∞
n=1 with xn  yn for all n,

lim xn  lim yn .
n→∞ n→∞
Continuous Preferences

Definition (continuous preference relation)


A preference relation  on X is continuous iff for any sequence
{(xn , yn )}∞
n=1 with xn  yn for all n,

lim xn  lim yn .
n→∞ n→∞

Theorem
A continuous, rational preference relation  on X ⊆ Rn can be
represented by a continuous utility function u : X → R.
Outline

1 Choice Theory
Preferences
Utility functions
Properties of preferences
Behavioral critiques
Comparative statics
An application: Consumer Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Reasons for restricting preferences

Analytical tractability often demands restricting “allowable”


preferences
Some restrictions are mathematical conveniences and cannot
be empirically falsified (e.g., continuity)
Some hold broadly (e.g., more is good–monotonicity.)
Some require situational justification
Reasons for restricting preferences

Analytical tractability often demands restricting “allowable”


preferences
Some restrictions are mathematical conveniences and cannot
be empirically falsified (e.g., continuity)
Some hold broadly (e.g., more is good–monotonicity.)
Some require situational justification

Restrictions on preferences imply restrictions on utility functions.


Reasons for restricting preferences

Analytical tractability often demands restricting “allowable”


preferences
Some restrictions are mathematical conveniences and cannot
be empirically falsified (e.g., continuity)
Some hold broadly (e.g., more is good–monotonicity.)
Some require situational justification

Restrictions on preferences imply restrictions on utility functions.


And perhaps more relevant, restrictions on utility functions imply
restrictions on preferences.
Reasons for restricting preferences

Analytical tractability often demands restricting “allowable”


preferences
Some restrictions are mathematical conveniences and cannot
be empirically falsified (e.g., continuity)
Some hold broadly (e.g., more is good–monotonicity.)
Some require situational justification

Restrictions on preferences imply restrictions on utility functions.


And perhaps more relevant, restrictions on utility functions imply
restrictions on preferences.
Assumptions for the rest of this section

1  is rational (i.e., complete and transitive).


2 For simplicity, we assume preferences over X ⊆ Rn .
Interpretation and notation

We are considering a choice from the set X ⊆ Rn

How can we interpret this?


Interpretation and notation

We are considering a choice from the set X ⊆ Rn

How can we interpret this?

What then is a element x ∈ X?


Interpretation and notation

We are considering a choice from the set X ⊆ Rn

How can we interpret this?

What then is a element x ∈ X?

What might we mean when x, y ∈ X and we say x > y?


Interpretation and notation

We are considering a choice from the set X ⊆ Rn

How can we interpret this?

What then is a element x ∈ X?

What might we mean when x, y ∈ X and we say x > y?


There are different conventions: For us, x is weakly greater than y
in every coordinate, and strictly greater in at least one coordinate.
Interpretation notation

Letting x, y ∈ Rn , we write

x ≥ y if and only if xi ≥ yi for all i.

x > y if and only if xi ≥ yi for all i and xi > yi for some i.

x >> y if and only if xi > yi for all i.


Properties of rational : Monotonicity

Definition (monotonicity)
 is monotone iff x > y =⇒ x  y. (N.B. MWG differs)
 is strictly monotone iff x > y =⇒ x  y.

In words: If you give me more of everything, I weakly (strictly)


prefer it.
Properties of rational : Nonsatiation

Definition (local non-satiation)


 is locally non-satiated iff for any y and ε > 0, there exists x
such that ||x − y|| ≤ ε and x  y.

In words: For every bundle y ∈ Rn there exists a close by bundle x


that is better. For example, a little bit more of everything might be
a bit better.

 is locally non-satiated iff u(·) has no local maxima in X


Properties of rational : Convexity I

Convex preferences capture the idea that agents like diversity


1 Satisfying in some ways: rather alternate between juice and
soda than have either one every day
2 Unsatisfying in others: rather have a glass of either one than a
mixture
3 Key question is granularity of goods being aggregated
Over time? What period?
Over what “bite size”?
Properties of rational : Convexity II

Definition (convexity)
 is convex iff x  y and x0  y together imply that

tx + (1 − t)x0  y for all t ∈ (0, 1).

Equivalently,  is convex iff the upper contour set of any y (i.e.,


{x ∈ X : x  y}) is a convex set.
 is strictly convex iff x  y and x0  y (with x 6= x0 ) together
imply that
tx + (1 − t)x0  y for all t ∈ (0, 1).

i.e., one never gets worse off by mixing goods


Examples illustrating convexity definition I

Indifference Curve

(y1,y2)

Good 2
Examples illustrating convexity definition II

(x1,x2)

(y1,y2) (x’1,x’2)

Good 2
Examples illustrating convexity definition III

Upper Contour Set

(y1,y2)

Good 2
Examples illustrating convexity definition IV

(x1,x2)

(y1,y2) (x’1,x’2)

Good 2
Examples illustrating convexity definition V

(x1,x2)
(y1,y2)

(x’1,x’2)

Good 2
Quasiconcavity

Definition (quasiconcavity)
f : X → R is quasiconcave iff for all x ∈ X, the upper contour set
of x 
UCS(x) ≡ ξ ∈ X : f (ξ) ≥ f (x)

 if f (ξ1 ) ≥ f (x) and f (ξ2 ) ≥ f (x), then


is a convex sets; i.e.,
f λξ1 + (1 − λ)ξ2 ≥ f (x) for all λ ∈ [0, 1].
f (·) is strictly quasiconcave iff for all x ∈ X, f (ξ1 ) ≥ f (x) and
f (ξ2 ) ≥ f (x) implies that f λξ1 + (1 − λ)ξ2 > f (x) for all
λ ∈ (0, 1).
Concavity and Quasiconcavity

A function f (x) that is concave is quasi-concave.

A function f (x) that is quasiconcave function need not be concave.


Examples: Concavity and Quasiconcavity I

x
Examples: Concavity and Quasiconcavity II

x
Examples: Concavity and Quasiconcavity III

Convex set

x
Examples: Concavity and Quasiconcavity IV

x
Examples: Concavity and Quasiconcavity V

x
Examples: Concavity and Quasiconcavity VI

x
Examples: Concavity and Quasiconcavity VII

x
Examples: Concavity and Quasiconcavity VIII

x
Examples: Concavity and Quasiconcavity IX

x
Convex preferences and quasiconcave utility functions

A preference relation  is (strictly) convex if and only if u is


(strictly) quasiconcave.
Example 1: Preferences

Suppose: (x1 , y1 )  (x2 , y2 ) if and only if


log(x1 ) + log(y1 ) ≥ log(x2 ) + log(y2 ).
10

0
0 2 4 6 8 10
Example 1: Preferences

Suppose: (x1 , y1 )  (x2 , y2 ) if and only if


log(x1 ) + log(y1 ) ≥ log(x2 ) + log(y2 ).
10

0
0 2 4 6 8 10
Example 1: Utility Function

Represent these preferences by u(x, y) = log(x) + log(y).


Example 1: Utility Function

Represent these preferences by u(x, y) = log(x) + log(y).


Example 1: Utility Function

Represent these preferences by u(x, y) = log(x) + log(y).


Example 2: Preferences

Suppose (x1 , y1 )  (x2 , y2 ) if and only if


log(x1 ) + log(y1 ) + max(x1 , y1 ) ≥ log(x2 ) + log(y2 ) + max(x2 , y2 ).
10

2 4 6 8 10
Example 2: Preferences

Suppose (x1 , y1 )  (x2 , y2 ) if and only if


log(x1 ) + log(y1 ) + max(x1 , y1 ) ≥ log(x2 ) + log(y2 ) + max(x2 , y2 ).
10

2 4 6 8 10
Example 2: Utility function

Represent these preferences by


u(x, y) = log(x) + log(y) + max(x, y).
Example 2: Utility function

Represent these preferences by


u(x, y) = log(x) + log(y) + max(x, y).
Example 2: Utility function

Represent these preferences by


u(x, y) = log(x) + log(y) + max(x, y).
Properties of rational : Homotheticity

Definition (homotheticity)
 is homothetic iff for all x, y, and all λ > 0,

x  y ⇐⇒ λx  λy.

Definition (Homogeneity of degree k)


A utility function u : Rn → R is homogeneous of degree k if and
only if for all x ∈ Rn and all λ > 0, u(λx) = λk u(x).
Properties of rational : Homotheticity

Definition (homotheticity)
 is homothetic iff for all x, y, and all λ > 0,

x  y ⇐⇒ λx  λy.

Definition (Homogeneity of degree k)


A utility function u : Rn → R is homogeneous of degree k if and
only if for all x ∈ Rn and all λ > 0, u(λx) = λk u(x).

Continuous, strictly monotone  is homothetic iff it can be


represented by a utility function that is
Properties of rational : Homotheticity

Definition (homotheticity)
 is homothetic iff for all x, y, and all λ > 0,

x  y ⇐⇒ λx  λy.

Definition (Homogeneity of degree k)


A utility function u : Rn → R is homogeneous of degree k if and
only if for all x ∈ Rn and all λ > 0, u(λx) = λk u(x).

Continuous, strictly monotone  is homothetic iff it can be


represented by a utility function that is homogeneous of degree one
(note it can also be represented by utility functions that aren’t)
Homogeneity of degree 1 and linearity

Suppose x ∈ R and a function f (x) is homogeneous of degree 1,


then f (x) = bx for some constant b. The function is linear in
terms of the calculus notion of linearity.
Homogeneity of degree 1 and linearity

Suppose x ∈ R and a function f (x) is homogeneous of degree 1,


then f (x) = bx for some constant b. The function is linear in
terms of the calculus notion of linearity.

In linear algebra, we use a different notion of linearity. A function


f (x) is defined to be linear (sometimes called a linear map) if:

f (a + b) = f (a) + f (b).

Thus any linear function is homogeneous of degree 1.

If x ∈ Rn for n > 1, a function f (x) that is homogeneous of


degree 1 need not be linear.
e.g. f (x1 , x2 ) = min(x1 , x2 )
Implications for utility representation

Property of  Property of u(·)


Monotone Nondecreasing
Strictly monotone Increasing
Locally non-satiated Has no local maxima in X
Convex Quasiconcave
Strictly convex Strictly quasiconcave
Implications for utility representation

Property of  Property of u(·)


Monotone Nondecreasing
Strictly monotone Increasing
Locally non-satiated Has no local maxima in X
Convex Quasiconcave
Strictly convex Strictly quasiconcave
Homothetic, continuous, can be represented by a
strictly monotone utility function that is
Homogeneous of degree 1

Why should you care?


Properties of rational : Separability I

Suppose rational  over X × Y ⊆ Rp+q


First p goods form some “group” x ∈ X ⊆ Rp
Other goods y ∈ Y ⊆ Rq

Separable preferences
“Preferences over X do not depend on y” means that

(x0 , y1 )  (x, y1 ) ⇐⇒ (x0 , y2 )  (x, y2 )

for all x, x0 ∈ X and all y1 , y2 ∈ Y .


Note the definition is not symmetric in X and Y .

The critical assumption for empirical analysis of preferences


Properties of rational : Separability II

Example
X = {wine, beer} and Y = {cheese, crisps} with strict preference
ranking
1 (wine, cheese) 
2 (wine, crisps) 
3 (beer, crisps) 
4 (beer, cheese).
Utility representation of separable preferences: theorem

Theorem
Suppose  on X × Y is represented by u(x, y). Then preferences
over X do not depend on y iff there exist functions v : X → R and
U : R × Y → R such that
1 U (·, ·) is increasing in its first argument, and
2 u(x, y) = U (v(x), y) for all (x, y).
Utility representation of separable preferences: example

Example
Preferences over beverages do not depend on your snack, and are
represented by u(·, ·), where

u(wine, cheese) =4
u(wine, crisps) =3
u(beer, crisps) =2
u(beer, cheese) = 1.
Utility representation of separable preferences: example

Example
Preferences over beverages do not depend on your snack, and are
represented by u(·, ·), where

u(wine, cheese) =4
u(wine, crisps) =3
u(beer, crisps) =2
u(beer, cheese) = 1.

Let v(wine) ≡ 3 and v(beer) ≡ 2.


Utility representation of separable preferences: example

Example
Preferences over beverages do not depend on your snack, and are
represented by u(·, ·), where

u(wine, cheese) =4 And let U (3, cheese) ≡4


u(wine, crisps) =3 U (3, crisps) ≡3
u(beer, crisps) =2 U (2, crisps) ≡2
u(beer, cheese) = 1. U (2, cheese) ≡ 1.

Let v(wine) ≡ 3 and v(beer) ≡ 2.


Utility representation of separable preferences: example

Example
Preferences over beverages do not depend on your snack, and are
represented by u(·, ·), where

u(wine, cheese) =4 And let U (3, cheese) ≡4


u(wine, crisps) =3 U (3, crisps) ≡3
u(beer, crisps) =2 U (2, crisps) ≡2
u(beer, cheese) = 1. U (2, cheese) ≡ 1.

Let v(wine) ≡ 3 and v(beer) ≡ 2.


Thus
1 U (·, ·) is increasing in its first argument, and
2 u(x, y) = U (v(x), y) for all (x, y).
Properties of rational : Quasilinearity

Suppose rational  over X ≡ R × Y


First good is the numeraire (a.k.a. “good zero” or “good
one,” confusingly): think money
Other goods general; need not be in Rn
Properties of rational : Quasilinearity

Theorem
Suppose rational  on X ≡ R × Y satisfies the “numeraire
properties”:
1 Good 1 is valuable: (t, y)  (t0 , y) ⇐⇒ t ≥ t0 for all y;
2 Compensation is possible: For every y, y 0 ∈ Y , there exists
some t ∈ R such that (0, y) ∼ (t, y 0 );
3 No wealth effects: If (t, y)  (t0 , y 0 ), then for all d ∈ R,
(t + d, y)  (t0 + d, y 0 ).
Then there exists a utility function representing  of the form
u(t, y) = t + v(y) for some v : Y → R. (Note it can also be
represented by utility functions that aren’t of this form.)
Conversely, any  on X = R × Y represented by a utility function
of the form u(t, y) = t + v(y) satisfies the above properties.
Outline

1 Choice Theory
Preferences
Utility functions
Properties of preferences
Behavioral critiques
Comparative statics
An application: Consumer Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Problems with rational choice

Rational choice theory plays a central role in most tools of


economic analysis

But. . . significant research calls into question underlying


assumptions, identifying and explaining deviations using
Psychology
Sociology
Cognitive neuroscience (“neuroeconomics”)
Context-dependent choice

Choices appear to be highly situational, depending on


1 Other available options
2 Way that options are “framed”
3 Social context/emotional state

Numerous research projects consider these effects in real-world and


laboratory settings
e.g., Ariely (2003, QJE); Madrian and Shea (2001, QJE).
Non-considered choice

Rational choice theory depends on a considered comparison of


options
Pairwise comparison
Utility maximization

Many actual choices appear to be made using


1 Intuitive reasoning
2 Heuristics
3 Instinctive desire
Policy

Sharp predictions based on rational agents.

Behavioral critiques call these predictions into question.


Policy

Sharp predictions based on rational agents.

Behavioral critiques call these predictions into question.

Could an alternative model/approach do better?


Outline

1 Choice Theory
Preferences
Utility functions
Properties of preferences
Behavioral critiques
Comparative statics
An application: Consumer Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Comparative statics

Comparative statics is the study of how endogenous variables


respond to changes in exogenous variables
Endogenous variables are typically set by
1 Maximization, or
2 Equilibrium
Comparative statics

Comparative statics is the study of how endogenous variables


respond to changes in exogenous variables
Endogenous variables are typically set by
1 Maximization, or
2 Equilibrium
Often we can characterize a maximization problem as a system of
equations (like an equilibrium)
Typically we do this using FOCs
Key comparative statics tool is the Implicit Function Theorem
Runs into lots of problems with continuity, smoothness,
nonconvexity, . . .
Comparative statics tools

We will discuss (and use):


1 Envelope Theorem

2 Implicit Function Theorem

Need different tools without continuity, smoothness, . . .


(but we won’t cover this):
1 Topkis’ Theorem

2 Monotone Selection Theorem


Envelope Theorem

The ET and IFT tell us about the derivatives of different objects


with respect to the parameters of the problem (i.e., exogenous
variables):
Envelope Theorems consider value function

Implicit Function Theorem considers choice function


Envelope Theorem

A simple Envelope Theorem:

v(q) = max f (x, q)


x
= f x∗ (q), q


d ∂ ∂  ∂ ∗
f x∗ (q), q + f x∗ (q), q

v(q) = x (q)
dq ∂q |∂x {z } ∂q
=0 by FOC

f x∗ (q), q

=
∂q

Think of the ET as an application of the chain rule and then FOCs


A more complete Envelope Theorem

Theorem (Envelope Theorem)


Consider a constrained optimization problem v(θ) = maxx f (x, θ)
such that g1 (x, θ) ≥ 0, . . . , gK (x, θ) ≥ 0.
A more complete Envelope Theorem

Theorem (Envelope Theorem)


Consider a constrained optimization problem v(θ) = maxx f (x, θ)
such that g1 (x, θ) ≥ 0, . . . , gK (x, θ) ≥ 0.
Comparative statics on the value function are given by:
K
∂v ∂f X ∂gk ∂L
= + λk =
∂θi ∂θi x∗ ∂θi x∗ ∂θi x∗
k=1
P
(for Lagrangian L(x, θ, λ) ≡ f (x, θ) + k λk gk (x, θ)) for all θ
such that the set of binding constraints does not change in an
open neighborhood.

Roughly, the derivative of the value function is the derivative of


the Lagrangian
Example: Cost Minimization Problem

Single-output cost minimization problem

min w · z such that f (z) ≥ q.


z∈Rm
+


L(q, w, λ, µ) ≡ −w · z + λ f (z) − q + µ · z

Applying Kuhn-Tucker here gives


∂f (z ∗ )
λ ≤ wi with equality if zi∗ > 0
∂zi
The ET applied to c(q, w) ≡ minz∈Rm
+ , f (z)≥q
w · z gives
Example: Cost Minimization Problem

Single-output cost minimization problem

min w · z such that f (z) ≥ q.


z∈Rm
+


L(q, w, λ, µ) ≡ −w · z + λ f (z) − q + µ · z

Applying Kuhn-Tucker here gives


∂f (z ∗ )
λ ≤ wi with equality if zi∗ > 0
∂zi
The ET applied to c(q, w) ≡ minz∈Rm
+ , f (z)≥q
w · z gives
∂c(q, w)

∂q
Example: Cost Minimization Problem

−c(q, w) = L(q, w, λ, µ)|z=z ∗ = −w · z ∗ + λ f (z ∗ ) − q + µ · z ∗




m
d − c(q, w) X ∂zi∗ ∂z ∗ ∂z ∗
= −λ + −wi + λfzi∗ (z ∗ ) i + µi i
dq ∂q ∂q ∂q
i=1
m
X  ∂zi∗
= −λ + −wi + λfzi∗ (z ∗ ) + µi
∂q
i=1
Lets look at the first order condition of the Lagrangian in zi :

−wi + λfzi (z) + µi = 0


Substituting this into the above equation:

dc(q, w)

dq
The Implicit Function Theorem I

A simple, general maximization problem

X ∗ (t) = argmax F (x, t)


x∈X

where F : X × T → R and X × T ⊆ R2 .
Suppose:
1 Smoothness: F is twice continuously differentiable
2 Convex choice set: X is convex
3 00 < 0
Strictly concave objective (in choice variable): Fxx
(together with convexity of X, this ensures a unique
maximizer)
4 Interiority: X ∗ (t) is in the interior of X for all t (which means
the standard FOC must hold)
The Implicit Function Theorem II

Under these assumptions there is a unique maximizer: |X ∗ (t)| = 1.


The first-order condition says the unique maximizer satisfies

Fx0 x(t), t = 0


Taking the derivative in t:

dFx0 x(t), t ∂Fx0 x(t), t ∂x(t) ∂Fx0 x(t), t


  
= + = 0.
dt ∂x(t) ∂t ∂t
So,
00 x(t), t

0 Fxt
x (t) = − 00 
Fxx x(t), t
Note by strict concavity, the denominator is negative, so x0 (t) and
00 x(t), t have the same sign
the cross-partial Fxt
Illustrating the Implicit Function Theorem

FOC: Fx0 x(t), t = 0




00 > 0 Thus t
Suppose Fxt high > tlow =⇒
Illustrating the Implicit Function Theorem

FOC: Fx0 x(t), t = 0




00 > 0 Thus t
Suppose Fxt 0 0
high > tlow =⇒ Fx (x, thigh ) > Fx (x, tlow )
Illustrating the Implicit Function Theorem

FOC: Fx0 x(t), t = 0




00 > 0 Thus t
Suppose Fxt 0 0
high > tlow =⇒ Fx (x, thigh ) > Fx (x, tlow )

x
Fx0 (·, tlow )

x(tlow )

Fx0 (·, thigh )


x

x(thigh )
Illustrating the Implicit Function Theorem

FOC: Fx0 x(t), t = 0




00 > 0 Thus t
Suppose Fxt 0 0
high > tlow =⇒ Fx (x, thigh ) > Fx (x, tlow )

F (·, tlow )
x
Fx0 (·, tlow )
x
x(tlow ) x(tlow )
F (·, thigh )
Fx0 (·, thigh )
x

x
x(thigh ) x(thigh )
Intuition for the Implicit Function Theorem

00 ≥ 0, an increase in x is more valuable when the


When Fxt
parameter t is higher
In a sense, x and t are complements; we therefore expect that an
increase in t results in an increase in the optimal choice of x
Intuition for the Implicit Function Theorem

00 ≥ 0, an increase in x is more valuable when the


When Fxt
parameter t is higher
In a sense, x and t are complements; we therefore expect that an
increase in t results in an increase in the optimal choice of x

This intuition should carry through without all our assumptions


Monotone comparative statics lead to the same conclusion
without smoothness of F or strict concavity of F in x.
Outline

1 Choice Theory
Preferences
Utility functions
Properties of preferences
Behavioral critiques
Comparative statics
An application: Consumer Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


The consumer problem

Utility Maximization Problem

max u(x) such that p·x ≤w


x∈Rn
+
|{z}
Expenses

where p are the prices of goods and w is the consumer’s “wealth.”

This type of choice set is a budget set

B(p, w) ≡ {x ∈ Rn+ : p · x ≤ w}

Can then write our problem as:

max u(x)
x∈B(p,w)
Illustrating the Utility Maximization Problem

Suppose there are two goods (X = R2 ), what are the best choices
of x given B(p, w)? When is there a unique best choice?
Assumptions underlying the UMP

Note that
Utility function is general (but assumed to exist—a restriction
of preferences)
Choice set defined by linear budget constraint
Consumers are price takers
Prices are linear
Perfect information: prices are all known
Finite number of goods
Goods are described by quantity and price
Goods are divisible
Goods may be time- or situation-dependent
Perfect information: goods are all well understood
Aside: Marshall

Cambridge mathematician who became Professor of Political


Economy at Cambridge
Utility maximization problem

The consumer’s Marshallian demand is given by correspondence


x : Rn × R ⇒ Rn+

x(p, w) ≡ argmax u(x) ≡ argmax u(x)


x∈Rn
+ : p·x≤w x∈B(p,w)

Rn+ :

= x∈ p · x ≤ w and u(x) = v(p, w)

Resulting indirect utility function is given by

v(p, w) ≡ sup u(x) ≡ sup u(x)


x∈Rn
+ : p·x≤w x∈B(p,w)
Properties of Marshallian demand and indirect utility

Theorem
v(p, w) and x(p, w) are homogeneous of degree zero. That is, for
all p, w, and λ > 0,

v(λp, λw) = v(p, w) and x(λp, λw) = x(p, w).

These are “no money illusion” conditions

Proof.
B(λp, λw) = B(p, w), so consumers are solving the same
problem.
Implications of restrictions on preferences: convexity I

Theorem
If preferences are convex, then x(p, w) is a convex set for every
p  0 and w ≥ 0.

Theorem
If preferences are strictly convex, then x(p, w) is single-valued for
every p  0 and w ≥ 0.
Implications of restrictions on preferences: convexity II
Implications of restrictions on preferences: non-satiation I

Definition (Walras’ Law)


p · x = w for every p  0, w ≥ 0, and x ∈ x(p, w).

Theorem
If preferences are locally non-satiated, then Walras’ Law holds.

This allows us to replace the inequality constraint in the UMP with


an equality constraint
Implications of restrictions on preferences: non-satiation II

Proof.
Suppose that p · x < w for some
x ∈ x(p, w). Then there exists
some x0 sufficiently close to x
with x0  x and p · x0 < w,
which contradicts the fact that
x ∈ x(p, w). Thus p · x = w.
Solving for Marshallian demand I

Suppose the utility function is differentiable


This is an ungrounded assumption
However, differentiability can not be falsified by any finite
data set
Also, utility functions are robust to monotone transformations
We may be able to use Kuhn-Tucker to “solve” the UMP:
Utility Maximization Problem

max u(x) such that p · x ≤ w


x∈Rn
+

gives the Lagrangian

L(x, λ, µ, p, w) ≡ u(x) + λ(w − p · x) + µ · x.


Solving for Marshallian demand II

1 First order conditions:

u0i (x∗ ) = λpi − µi for all i

2 Complementary slackness:

λ(w − p · x∗ ) = 0
µi x∗i = 0 for all i

3 Non-negativity:

λ ≥ 0 and µi ≥ 0 for all i

4 Original constraints p · x∗ ≤ w and x∗i ≥ 0 for all i


We can solve this system of equations for certain functional forms
of u(·)
The power (and limitations) of Kuhn-Tucker

Kuhn-Tucker provides conditions on (x, λ, µ) given (p, w):


1 First order conditions
2 Complementary slackness
3 Non-negativity
4 (Original constraints)

Kuhn-Tucker tells us that if x∗ is a solution to the UMP, there


exist some (λ, µ) such that these conditions hold
The power (and limitations) of Kuhn-Tucker

Kuhn-Tucker provides conditions on (x, λ, µ) given (p, w):


1 First order conditions
2 Complementary slackness
3 Non-negativity
4 (Original constraints)

Kuhn-Tucker tells us that if x∗ is a solution to the UMP, there


exist some (λ, µ) such that these conditions hold; however:
These are only necessary conditions; there may be (x, λ, µ)
that satisfy Kuhn-Tucker conditions but do not solve UMP
If u(·) is concave, conditions are necessary and sufficient
When are Kuhn-Tucker conditions sufficient?

Kuhn-Tucker conditions are necessary and sufficient for a solution


(assuming differentiability) as long as we have a “convex problem”:

1 The constraint set is convex


If each constraint gives a convex set, the intersection is a
convex set

The set x : gk (x, θ) ≥ 0 is convex as long as gk (·, θ) is a
quasiconcave function of x

2 The objective function is concave


If we only know the objective is quasiconcave, there are other
conditions that ensure Kuhn-Tucker is sufficient
Intuition from Kuhn-Tucker conditions I

Recall (evaluating at the optimum, and for all i):


FOC u0i (x) = λpi − µi
CS λ(w − p · x) = 0 and µi xi = 0
NN λ ≥ 0 and µi ≥ 0
Orig p · x ≤ w and xi ≥ 0
We can summarize as

u0i (x) ≤ λpi with equality if xi > 0

And therefore if xj > 0 and xk > 0,


∂u
pj ∂xj
= ∂u
≡ MRSjk
pk ∂xk
Intuition from Kuhn-Tucker conditions II

The MRS is the (negative) slope of the indifference curve


Price ratio is the (negative) slope of the budget line

x2
6 @
@
@
@
@
@ x∗  Du(x∗ )
q
@
p
@
@
@
@
@
@
@
@ - x1
Intuition from Kuhn-Tucker conditions III

Recall the Envelope Theorem tells us the derivative of the value


function in a parameter is the derivative of the Lagrangian:
Value function (indirect utility)

v(p, w) ≡ sup u(x)


x∈B(p,w)

Lagrangian

L ≡ u(x) + λ(w − p · x) + µ · x
∂v
By the Envelope Theorem, ∂w = λ; i.e., the Lagrange multiplier λ
is the “shadow value of wealth” measured in terms of utility
Intuition from Kuhn-Tucker conditions IV

Given our envelope result, we can interpret our earlier condition

∂u
= λpi if xi > 0
∂xi

as

∂u ∂v
= pi if xi > 0
∂xi ∂w
where each side gives the marginal utility from an extra unit of xi
LHS directly
RHS through the wealth we could get by selling it
MRS and separable utility

Recall that if xj > 0 and xk > 0,


∂u
∂xj
MRSjk ≡ ∂u
∂xk

does not depend on λ; however it typically depends on x1 , . . . , xn


MRS and separable utility

Recall that if xj > 0 and xk > 0,


∂u
∂xj
MRSjk ≡ ∂u
∂xk

does not depend on λ; however it typically depends on x1 , . . . , xn


Suppose choice from X × Y where preferences over X do not
depend on y

Recall that u(x, y) = U v(x), y for some U (·, ·) and v(·)
∂u 0 0
 ∂v ∂u
 ∂v
∂xj = U1 v(x), y ∂xj and ∂xk = U1 v(x), y ∂xk
∂v ∂v
MRSjk = ∂xj / ∂xk does not depend on y
Separability allows empirical work without worrying about y
Recap: Utility Maximization Problem

Basic problem: How much of different goods will a consumer buy?


Suppose preferences are rational and satisfy continuity and
local non-satiation.
Can represent the consumer’s problem as a constrained
optimization problem.
There exists a continuous and differentiable utility function
representing these preferences.
Any such utility function
Recap: Utility Maximization Problem

Basic problem: How much of different goods will a consumer buy?


Suppose preferences are rational and satisfy continuity and
local non-satiation.
Can represent the consumer’s problem as a constrained
optimization problem.
There exists a continuous and differentiable utility function
representing these preferences.
Any such utility function has no local maxima.
Recap: Utility Maximization Problem

Basic problem: How much of different goods will a consumer buy?


Suppose preferences are rational and satisfy continuity and
local non-satiation.
Can represent the consumer’s problem as a constrained
optimization problem.
There exists a continuous and differentiable utility function
representing these preferences.
Any such utility function has no local maxima.

The Utility Maximization problem


Choose a vector x ∈ Rn+ describing how much we want to
consume of the n available goods.
To maximize utility u(x).
Selecting from bundles that are affordable: p · x ≤ w
Recap: Utility Maximization Problem I

The Utility Maximization problem

max u(x) such that p · x ≤ w


x∈Rn
+
Recap: Utility Maximization Problem I

The Utility Maximization problem

max u(x) such that p · x ≤ w


x∈Rn
+

And this gives the Lagrangian:

L(x, λ, µ, p, w) ≡ u(x) + λ(w − p · x) + µ · x.


Recap: Utility Maximization Problem II

Now solve the Lagrangian.


First differentiate with respect to xi to get first order
conditions:
Recap: Utility Maximization Problem II

Now solve the Lagrangian.


First differentiate with respect to xi to get first order
conditions:

∂L(x, λ, µ, p, w) ∂u(x)
= − λpi + µi = 0 for all i
∂xi ∂xi
Any solution to the consumer’s problem must satisfy these
conditions.

But which constraints bind?


Recap: Utility Maximization Problem II

Now solve the Lagrangian.


First differentiate with respect to xi to get first order
conditions:

∂L(x, λ, µ, p, w) ∂u(x)
= − λpi + µi = 0 for all i
∂xi ∂xi
Any solution to the consumer’s problem must satisfy these
conditions.

But which constraints bind? Using local non-satiation, apply


Walras’ law.
Recap: Utility Maximization Problem II

Now solve the Lagrangian.


First differentiate with respect to xi to get first order
conditions:

∂L(x, λ, µ, p, w) ∂u(x)
= − λpi + µi = 0 for all i
∂xi ∂xi
Any solution to the consumer’s problem must satisfy these
conditions.

But which constraints bind? Using local non-satiation, apply


Walras’ law.

This tells us that w = p · x.


Recap: Utility Maximization Problem II

Now solve the Lagrangian.


First differentiate with respect to xi to get first order
conditions:

∂L(x, λ, µ, p, w) ∂u(x)
= − λpi + µi = 0 for all i
∂xi ∂xi
Any solution to the consumer’s problem must satisfy these
conditions.

But which constraints bind? Using local non-satiation, apply


Walras’ law.

This tells us that w = p · x. We also know λ > 0. Why?


Recap: Utility Maximization Problem II

Answer 1:
The first order condition in xi is:
∂u(x)
= λpi − µi .
∂xi
If λ = 0 then ∂u(x)/∂xi ≤ 0 for all i.
But then there is a local maxima.
Recap: Utility Maximization Problem II

Answer 1:
The first order condition in xi is:
∂u(x)
= λpi − µi .
∂xi
If λ = 0 then ∂u(x)/∂xi ≤ 0 for all i.
But then there is a local maxima.

Answer 2:
As there are no local maxima utility must be increasing locally
in some xi .
So ∂v/∂w > 0.
But by the Envelope Theorem, ∂v/∂w = ∂L/∂w = λ.
Recap: Utility Maximization Problem III

What about the non-negativity constraints? Do they bind


too?
Recap: Utility Maximization Problem III

What about the non-negativity constraints? Do they bind


too?
Know from complementary slackness that µi xi = 0.
So, if xi > 0 then µi = 0.
Recap: Utility Maximization Problem III

What about the non-negativity constraints? Do they bind


too?
Know from complementary slackness that µi xi = 0.
So, if xi > 0 then µi = 0.
From the first order conditions, if xi > 0 and xj > 0

∂u(x) ∂u(x)
= λpi and = λpj
∂xi ∂xj
Recap: Utility Maximization Problem IV

Combining

∂u(x) ∂u(x)
= λpi and = λpj ,
∂xi ∂xj
we get
∂u(x)
∂xi pi
∂u(x)
= .
pj
∂xj
Recap: Utility Maximization Problem IV

Combining

∂u(x) ∂u(x)
= λpi and = λpj ,
∂xi ∂xj
we get
∂u(x)
∂xi pi
∂u(x)
= .
pj
∂xj

Left hand side is the marginal rate of substitution.


Right hand side is the price ratio.
Why we need another “problem”

We would like to characterize “important” properties of


Marshallian demand x(·, ·) and indirect utility v(·, ·)
For instance, does demand for a good always increase when
its price decreases?
Difficult because in the UMP parameters enter feasible set
rather than objective
Why we need another “problem”

We would like to characterize “important” properties of


Marshallian demand x(·, ·) and indirect utility v(·, ·)
For instance, does demand for a good always increase when
its price decreases?
Difficult because in the UMP parameters enter feasible set
rather than objective

Consider a price increase for one good (apples)


1 Substitution effect: Apples are now relatively more expensive
than bananas, so I buy fewer apples
2 Wealth effect: I feel poorer, so I buy (more? fewer?)
apples
Wealth effect and substitution effects could go in opposite
directions =⇒ can’t easily sign the change in consumption
Isolating the substitution effect

We can isolate the substitution effect by “compensating” the


consumer so that her maximized utility does not change
If maximized utility doesn’t change, the consumer can’t feel richer
or poorer; demand changes can therefore be attributed entirely to
the substitution effect
Isolating the substitution effect

We can isolate the substitution effect by “compensating” the


consumer so that her maximized utility does not change
If maximized utility doesn’t change, the consumer can’t feel richer
or poorer; demand changes can therefore be attributed entirely to
the substitution effect

Expenditure Minimization Problem

min p · x such that u(x) ≥ ū.


x∈Rn
+

i.e., find the cheapest bundle at prices p that yield utility at least ū
Illustrating the Expenditure Minimization Problem
Aside: Hicks

Lecturer at Cambridge, 1935-38.


Expenditure minimization problem

The consumer’s Hicksian demand is given by correspondence


h : Rn × R ⇒ Rn

h(p, ū) ≡ argmin p·x


x∈Rn
+ : u(x)≥ū

= {x ∈ Rn+ : u(x) ≥ ū and p · x = e(p, ū)}

Resulting expenditure function is given by

e(p, ū) ≡ min p·x


x∈Rn
+ : u(x)≥ū
Illustrating Hicksian demand
Recap

x(p, w) bundles that maximize utility given


budget constraint

v(p, w) utility generated by optimal bundles


given budget constraint

h(p, u) bundles than minimize expenditure


for given utility

e(p, u) minimum required expenditure


to obtain given utility
Relating Hicksian and Marshallian demand I

Theorem (“Same problem” identities)


Suppose u(·) is a utility function representing a continuous and
locally non-satiated preference relation  on Rn+ . Then for any
p  0 and w ≥ 0,

1 h p, v(p, w) = x(p, w),

2 e p, v(p, w) = w;

and for any ū ≥ u(0),



3 x p, e(p, ū) = h(p, ū), and

4 v p, e(p, ū) = ū.
Relating Hicksian and Marshallian demand II

These say that UMP and EMP are fundamentally solving the same
problem, so:
If the utility you can get with wealth w is v(p, w). . .
To achieve utility v(p, w) will cost at least w
You will buy the same bundle whether you have w to spend, or
you are trying to achieve utility v(p, w)

If it costs e(p, ū) to achieve utility ū. . .


Given wealth e(p, ū) you will achieve utility at most ū
You will buy the same bundle whether you have e(p, ū) to
spend, or you are trying to achieve utility ū
Relating (changes in) Hicksian and Marshallian demand I

Assuming differentiability and hence single-valuedness, we can


differentiate the ith row of the identity

h(p, ū) = x p, e(p, ū)

in pj to get

∂hi ∂xi ∂xi ∂e


= +
∂pj ∂pj ∂e ∂pj
Note that
e(p, ū) = p · h(p, ū),
so
∂e(p, ū)
= hj (p, ū) = xj
∂pj
| {z }
Shepherd’s Lemma
The Slutsky equation I

Slutsky equation

∂xi (p, w) ∂hi p, u(x(p, w)) ∂xi (p, w)
= − xj (p, w)
∂pj ∂pj ∂w
| {z } | {z } | {z }
total effect substitution effect wealth effect

for all i and j.

Sometimes known as the Hicks decomposition (just to confuse


you, something different is known as the Slutsky decomposition!).
The Slutsky equation II

Setting i = j, we can decompose the effect of an increase in pi



∂xi (p, w) ∂hi p, u(x(p, w)) ∂xi (p, w)
= − xi (p, w)
∂pi ∂pi ∂w

An “own-price” increase. . .
1 Encourages consumer to substitute away from good i
∂hi
∂pi ≤0

2 Makes consumer poorer, which affects consumption of good i


in some indeterminate way
∂xi
Sign of ∂w depends on preferences
Illustrating wealth and substitution effects

Following a decrease in the price of the first good. . .


Substitution effect moves from x to h
Wealth effect moves from h to x0

Z
J
JZZ
J Z
J ZZ
J x Z
J Z
J Z
J Z
Z
J Z
JJ Z
Z -
Illustrating wealth and substitution effects

Following a decrease in the price of the first good. . .


Substitution effect moves from x to h
Wealth effect moves from h to x0

Z
J
JZZ
J Z
Z J ZZ
Z J x Z
ZJ Z
Z
0 Z
Z
JZ h(p , u)
J
Z
Z Z
J Z Z
JJ ZZ Z
Z -
Illustrating wealth and substitution effects

Following a decrease in the price of the first good. . .


Substitution effect moves from x to h
Wealth effect moves from h to x0

Z
J
JZZ
J Z
Z J ZZ x0
Z J x Z
ZJ Z
Z
0 Z
Z
JZ h(p , u)
J
Z
Z Z
J Z Z
JJ ZZ Z
Z -
Marshallian response to changes in wealth

Definition (Normal good)


Good i is a normal good if xi (p, w) is increasing in w.

Definition (Inferior good)


Good i is an inferior good if xi (p, w) is decreasing in w.
Graphing Marshallian response to changes in wealth

Engle curves show how Marshallian demand moves with


wealth (locus of {x, x0 , x00 , . . . } below)
In this example, both goods are normal (xi increases in w)

Z
Z
Z
Z Z
Z x00
Z Z
Z Z
Z Z
x 0 Z
Z Z Z
Z x Z Z
Z Z Z
Z Z Z
Z Z Z
Z
Z Z
Z Z -
Marshallian response to changes in own price

Definition (Regular good)


Good i is a regular good if xi (p, w) is decreasing in pi .

Definition (Giffen good)


Good i is a Giffen good if xi (p, w) is increasing in pi .

Potatoes during the Irish potato famine are the canonical example
(and probably weren’t actually Giffen goods)
Marshallian response to changes in own price

Definition (Regular good)


Good i is a regular good if xi (p, w) is decreasing in pi .

Definition (Giffen good)


Good i is a Giffen good if xi (p, w) is increasing in pi .

Potatoes during the Irish potato famine are the canonical example
(and probably weren’t actually Giffen goods)
∂xi ∂hi ∂xi
By the Slutsky equation (which gives ∂pi = ∂pi − ∂w xi for i = j)
Normal =⇒ regular
Giffen =⇒ inferior
Graphing Marshallian response to changes in own price

Offer curves show how Marshallian demand moves with price


In this example, good 1 is regular and good 2 is a gross
complement for good 1

Q
@
JQ
J@Q
J@QQ
00
J@ Q
J @ xQ 0 x
x
J @
Q
Q
J @ Q
Q
J @ Q
J @ Q
Q
J @ Q
J @ Q
Q -
Marshallian response to changes in other goods’ price

Definition (Gross substitute)


Good i is a gross substitute for good j if xi (p, w) is increasing in
pj .

Definition (Gross complement)


Good i is a gross complement for good j if xi (p, w) is decreasing
in pj .

Gross substitutability/complementarity is not necessarily symmetric


Hicksian response to changes in other goods’ price

Definition (Substitute)
Good i is a substitute for good j if hi (p, ū) is increasing in pj .

Definition (Complement)
Good i is a complement for good j if hi (p, ū) is decreasing in pj .

Substitutability/complementarity is symmetric

In a two-good world, the goods must be substitutes (why? )


Wrapping up: Conceptual

Agents (people, firms, etc) make choices from sets of


alternatives.
Wrapping up: Conceptual

Agents (people, firms, etc) make choices from sets of


alternatives.
Their preferences determine these choices.
Wrapping up: Conceptual

Agents (people, firms, etc) make choices from sets of


alternatives.
Their preferences determine these choices.
Can be represented by a utility function if and only if rational.
Wrapping up: Conceptual

Agents (people, firms, etc) make choices from sets of


alternatives.
Their preferences determine these choices.
Can be represented by a utility function if and only if rational.
Properties of some underlying preferences have equivalent
properties in utility space.
Wrapping up: Conceptual

Agents (people, firms, etc) make choices from sets of


alternatives.
Their preferences determine these choices.
Can be represented by a utility function if and only if rational.
Properties of some underlying preferences have equivalent
properties in utility space.
We work in utility space because constrained maximization
problems are easier than binary comparisons!
Wrapping up: Conceptual

Agents (people, firms, etc) make choices from sets of


alternatives.
Their preferences determine these choices.
Can be represented by a utility function if and only if rational.
Properties of some underlying preferences have equivalent
properties in utility space.
We work in utility space because constrained maximization
problems are easier than binary comparisons!
Seen that it helps us understand consumers’ decisions.
Wrapping up: Mathematical tools we’ve used

Constrained optimization problems.

Envelope theorems.

Implicit function theorem.


Wrapping up: Language I

Preferences
- Completeness
- Transitivity
- Rationality
- Reflexivity
- Symmetry
- Continuity
- Monotonicity
- Local non-satiation
- Convexity
- Separability
- Choice rule
- Numeraire Properties
Wrapping up: Language II

Utility representation
- Choice rule
- Revealed preference choice rule
- HARP / WARP
- Quasiconcavity
- Homogeneity of degree k
Consumer theory
- Budget set
- Marshallian Demand
- Hicksian Demand
- Indirect utility
- Expenditure function
- Substitution effect
Wrapping up: Language III

- Wealth effect
- Normal goods
- Inferior goods
- Regular goods
- Giffen goods
- Substitutes
- Complements
- Gross Substitutes
- Gross Complements
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Why study uncertainty?

So far we have covered individual decision-making under certainty


Goods well understood
Prices well known

In fact, decisions typically made in the face of an uncertain future


Workhorse model: objective risk
Subjective assessments of uncertainty
Behavioral critiques
von Neumann-Morganstern expected utility model

Simplifying assumptions include


Finite number of outcomes (“prizes”)

Objectively known probability distributions over prizes


(“lotteries”)

Complete and transitive preferences over lotteries


Other assumptions on preferences over lotteries (to be
discussed)
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Prizes and lotteries

Let X be the set of possible prizes (a.k.a. outcomes or


consequences)
Assume |X | = n < ∞
Since |X | < ∞, there must be a best outcome and a worst
outcome
Prizes and lotteries

Let X be the set of possible prizes (a.k.a. outcomes or


consequences)
Assume |X | = n < ∞
Since |X | < ∞, there must be a best outcome and a worst
outcome

A lottery is a probability distribution over prizes


p = (p1 , . . . , pn ) ∈ Rn+
The set of all lotteries is
n X o
∆(X ) ≡ p ∈ Rn+ : pi = 1 ,
i

called the n-dimensional simplex


Graphing the simplex ∆(X ) ⊆ R2

Suppose there are two prizes (|X | = 2)


The simplex ∆(X ) is the portion of the line p1 + p2 = 1 that
lies in the positive quadrant

p1

∆(X )

p2
Graphing the simplex ∆(X ) ⊆ R3

Suppose there are three prizes (|X | = 3)


The simplex ∆(X ) is the portion of the plane
p1 + p2 + p3 = 1 that lies in the positive orthant
p2

p3 p1
Graphing the simplex ∆(X ) ⊆ R3

Suppose there are three prizes (|X | = 3)


The simplex ∆(X ) is the portion of the plane
p1 + p2 + p3 = 1 that lies in the positive orthant
p2 p2

p3 p1 p3 p1
Convexity of the simplex I

Note that ∆(X ) is a convex set


If pi ≥ 0 and p0i ≥ 0, then αpi + (1 − α)p0i ≥ 0
If i pi = 1 and i p0i = 1, then i [αpi + (1 − α)p0i ] = 1
P P P

This is not surprising given that the simplex is a “triangle”

p2

p
αp + (1 − α)p0
p0
p3 p1
Convexity of the simplex II

We can view αp + (1 − α)p0 as a compound lottery


1 Choose between lotteries: Lottery p with probability α and
lottery p0 with probability (1 − α)
2 Resolve uncertainty in chosen lottery per p or p0

p
α

1−α
p0
Preferences over lotteries

A rational decision-maker has preferences over outcomes X

We consider preferences over lotteries ∆(X ) (note that from here


on,  refers to preferences over lotteries, not outcomes)

Expected utility theory relies on  satisfying


Completeness
Transitivity
Continuity (in a sense to be defined)
Independence (to be defined)
Continuity axiom

Definition (continuity)
A preference relation  over ∆(X ) is continuous iff for any pH ,
pM , and pL ∈ ∆(X ) such that pH  pM  pL , there exists some
α ∈ [0, 1] such that

αpH + (1 − α)pL ∼ pM .
An example I

Consider the following lotteries:


p $12 for sure.
p0 $15 with probability 1/2 and $10 with probability 1/2.
pm $100 with probability 1/2 and $0 with probability 1/2.
An example I

Consider the following lotteries:


p $12 for sure.
p0 $15 with probability 1/2 and $10 with probability 1/2.
pm $100 with probability 1/2 and $0 with probability 1/2.

Suppose pm  p0  p.

Continuity implies that there exists an α ∈ [0, 1] such that


αpm + (1 − α)p ∼ p0 . Reasonable?
An example I

Consider the following lotteries:


p $12 for sure.
p0 $15 with probability 1/2 and $10 with probability 1/2.
pm $100 with probability 1/2 and $0 with probability 1/2.

Suppose pm  p0  p.

Continuity implies that there exists an α ∈ [0, 1] such that


αpm + (1 − α)p ∼ p0 . Reasonable?

When might continuity not hold?


An example II

Same lotteries:
p $12 for sure.
p0 $15 with probability 1/2 and $10 with probability 1/2.
pm $100 with probability 1/2 and $0 with probability 1/2.

Suppose p  p0 .

For some α ∈ [0, 1] consider the following two compound lotteries:


c1 αp0 + (1 − α)pm .
c2 αp + (1 − α)pm .
An example II

Same lotteries:
p $12 for sure.
p0 $15 with probability 1/2 and $10 with probability 1/2.
pm $100 with probability 1/2 and $0 with probability 1/2.

Suppose p  p0 .

For some α ∈ [0, 1] consider the following two compound lotteries:


c1 αp0 + (1 − α)pm .
c2 αp + (1 − α)pm .

Would you expect c1 be preferred to c2?


An example III

Compound Lottery c1:


1/2 100

α 0
1/2
1/2 15
1-α
1/2 10
Compound Lottery c2:
1/2 100

α
1/2 0

1-α 1 12
Independence axiom

Definition (independence)
A preference relation  over ∆(X ) satisfies independence iff for
any p, p0 , and pm ∈ ∆(X ) and any α ∈ [0, 1], we have

p  p0
m
αp + (1 − α)pm  αp0 + (1 − α)pm .

i.e., if I prefer p to p0 , I also prefer the possibility of p to the


possibility of p0 , as long as the other possibility is the same (a
(1 − α) chance of pm ) in both cases
Independence sensible for choice under uncertainty

There is no counterpart in standard consumer theory; e.g.,


p = (2 cokes, 0 apples) and p0 = (0 cokes, 2 apples)
pm = (2 cokes, 2 apples)
1
α= 2

p p0
z }| { z }| {
(2 cokes, 0 apples)  (0 cokes, 2 apples)
Independence sensible for choice under uncertainty

There is no counterpart in standard consumer theory; e.g.,


p = (2 cokes, 0 apples) and p0 = (0 cokes, 2 apples)
pm = (2 cokes, 2 apples)
1
α= 2
There is no reason to conclude that
p p0
z }| { z }| {
(2 cokes, 0 apples)  (0 cokes, 2 apples)
m
(2 cokes, 1 apples)  (1 cokes, 2 apples)
| {z } | {z }
αp+(1−α)pm αp0 +(1−α)pm
Independence sensible for choice under uncertainty

There is no counterpart in standard consumer theory; e.g.,


p = (2 cokes, 0 apples) and p0 = (0 cokes, 2 apples)
pm = (2 cokes, 2 apples)
1
α= 2
There is no reason to conclude that
p p0
z }| { z }| {
(2 cokes, 0 apples)  (0 cokes, 2 apples)
m
(2 cokes, 1 apples)  (1 cokes, 2 apples)
| {z } | {z }
αp+(1−α)pm αp0 +(1−α)pm

Why the difference?


Example

p = (2 cokes, 0 apples) and p0 = (0 cokes, 2 apples)


pm = (2 cokes, 2 apples)
pc = (2 cokes, 1 apples) and p0c = (1 cokes, 2 apples).

Suppose you are very hungry and thirsty.


(i) First priority is fluids: p  p0
(ii) But after 1 drink, food becomes more desirable: p0c  pc
(iii) Now consider a 50:50 chance of p and pm Vs. a 50:50 chance
of p0 and pm .
Example

p = (2 cokes, 0 apples) and p0 = (0 cokes, 2 apples)


pm = (2 cokes, 2 apples)
pc = (2 cokes, 1 apples) and p0c = (1 cokes, 2 apples).

Suppose you are very hungry and thirsty.


(i) First priority is fluids: p  p0
(ii) But after 1 drink, food becomes more desirable: p0c  pc
(iii) Now consider a 50:50 chance of p and pm Vs. a 50:50 chance
of p0 and pm .
(iv) (ii) and (iii) are not the same. Prefer the 50:50 chance of p
and pm because it guarantees liquids. . .
Independence implies linear indifference curves

Independence implies linear indifference curves


Consider p ∼ p0
Let pm = p0
By the independence axiom,

αp + (1 − α)p0 ∼ αp0 + (1 − α)p0


∼ p0
∼p

p2

p
αp + (1 − α)p0
p0
p3 p1
Independence implies parallel indifference curves

Independence implies parallel indifference curves


Consider p ∼ p0
Let pm be some other point, and α some value in (0, 1)
By independence αp + (1 − α)pm ∼ αp0 + (1 − α)pm
αp + (1 − α)pm and αp0 + (1 − α)pm lie on a line parallel to
the indifference curve containing p and p0

p2

p0
pm
p3 p1
Independence implies parallel indifference curves

Independence implies parallel indifference curves


Consider p ∼ p0
Let pm be some other point, and α some value in (0, 1)
By independence αp + (1 − α)pm ∼ αp0 + (1 − α)pm
αp + (1 − α)pm and αp0 + (1 − α)pm lie on a line parallel to
the indifference curve containing p and p0

p2

p0
pm
p3 p1
Recap: The Simplex

p2

(1, 0, 0)
p3 p1
Recap: The Simplex

p2

(1/2, 1/2, 0)

p3 p1
Recap: The Simplex

p2

p3 p1
Recap: The Simplex

p2

(1/6, 1/6, 2/3)

p3 p1
Linear and parallel indifference curves

Suppose X = {x1 , x2 , x3 } and x1  x2  x3 .

Let p1 = (1, 0, 0), p2 = (0, 1, 0) and p3 = (0, 0, 1).

By continuity, there exists a γ ∈ [0, 1] such that


pc = γp1 + (1 − γ)p3 ∼ p2 .
Linear and parallel indifference curves

p2

p3 p1
Linear and parallel indifference curves

p2

γ 1-γ
p3 p1
Linear and parallel indifference curves

By independence for any q, q 0 , and qm ∈ ∆(X ) and any α ∈ [0, 1],


we have

q  q0
m
αq + (1 − α)qm  αq 0 + (1 − α)qm .

Suppose we then have q  q 0 and q 0  q, then

αq + (1 − α)qm ∼ αq 0 + (1 − α)qm .
Linear and parallel indifference curves

By independence for any q, q 0 , and qm ∈ ∆(X ) and any α ∈ [0, 1],


we have

q  q0
m
αq + (1 − α)qm  αq 0 + (1 − α)qm .

Suppose we then have q  q 0 and q 0  q, then

αq + (1 − α)qm ∼ αq 0 + (1 − α)qm .
Letting q = p2 , q 0 = pc and qm = p2 , as p2  pc and pc  p2 , we
have

αp2 + (1 − α)p2 = p2 ∼ αpc + (1 − α)p2


Linear and parallel indifference curves

p2

γ 1-γ
p3 p1
Linear and parallel indifference curves

Using independence again, as pc ∼ p2 :

αpc + (1 − α)p1 ∼ αp2 + (1 − α)p1


There is thus an indifference curve connecting these points.
Linear and parallel indifference curves

p2

α 1-α
p3 p1
Linear and parallel indifference curves

p2
α

1-α

α 1-α
p3 p1
Linear and parallel indifference curves

p2
α

1-α

α 1-α
p3 p1
Linear and parallel indifference curves

β
p2

1-β

1-β β
p3 p1
Linear and parallel indifference curves

p2

p3 p1
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


von Neumann-Morganstern utility functions

Definition (von Neumann-Morganstern utility function)


A utility function U : ∆(X ) → R is a vNM utility function iff there
exist numbers u1 , . . . , un ∈ R such that for every p ∈ ∆(X ),
n
X
U (p) = pi ui = p · ~u.
i=1

Can think of u1 , . . . , un as indexing preference over outcomes


Linearity of vNM utility functions I

Theorem
A utility function U : ∆(X ) → R is a vNM utility function iff it is
linear in probabilities, i.e.,

U αp + (1 − α)p0 = αU (p) + (1 − α)U (p0 )




for all p, p0 ∈ ∆(X ), and α ∈ [0, 1]


Linearity of vNM utility functions II

Proof Outline
vNM =⇒ linearity:
Suppose |X | = 3.
Consider p = (p1 , p2 , p3 ) and p0 = (p01 , p02 , p03 ).
Let ~u = (u1 , u2 , u3 )

U αp + (1 − α)p0 = αp + (1 − α)p0 · ~u
 

= (αp) · ~u + (1 − α)p0 · ~u


= α(p · ~u) + (1 − α)(p0 · ~u)


= αU (p) + (1 − α)U (p0 )
Linearity of vNM utility functions III

linearity =⇒ vNM:
Continue to let |X | = 3 and p = (p1 , p2 , p3 ).
We can write p as a compound lottery:
p = p1 (1, 0, 0) + p2 (0, 1, 0) + p3 (0, 0, 1) = p1 b1 + p2 b2 + p3 b3

By linearity:
 X
U p = p1 U b1 + p2 U b2 + p3 U b3 =
  
pi ui
i

for ui = U bi .

Expected utility representation and ordinality

If preferences  can be represented by a vNM utility function, we


say it is an “expected utility representation” of 
That is “U (·) is an expected utility representation of ” means
1 U (·) is a vNM utility function, and
2 U (·) represents 
Expected utility representation and ordinality

If preferences  can be represented by a vNM utility function, we


say it is an “expected utility representation” of 
That is “U (·) is an expected utility representation of ” means
1 U (·) is a vNM utility function, and
2 U (·) represents 

Linearity of vNM utility functions mean that expected utility


representation is not ordinal
Utility representation is robust to any increasing monotone
transformation
Expected utility representation is only robust to affine
(increasing linear) transformations
Exp. util. representation robust to affine transformation

Theorem
Suppose U : ∆(X ) → R is an expected utility representation of .
Then V : ∆(X ) → R is also an expected utility representation of 
iff there exist some a ∈ R and b ∈ R++ such that

V (p) = a + bU (p)

for all p ∈ ∆(X ).


Level sets of an expected utility function

A von Neumann-Morganstern utility function satisfies


n
X
U (p) = pi ui = p · ~u
i=1

for some ~u ∈ Rn

Indifference curves are therefore p · ~u = c for various c


Level sets of an expected utility function

A von Neumann-Morganstern utility function satisfies


n
X
U (p) = pi ui = p · ~u
i=1

for some ~u ∈ Rn

Indifference curves are therefore p · ~u = c for various c


Indifference curves are therefore
Straight lines (p · ~u = c is a plane that intercepts the simplex
in a line)
Level sets of an expected utility function

A von Neumann-Morganstern utility function satisfies


n
X
U (p) = pi ui = p · ~u
i=1

for some ~u ∈ Rn

Indifference curves are therefore p · ~u = c for various c


Indifference curves are therefore
Straight lines (p · ~u = c is a plane that intercepts the simplex
in a line)
Parallel (all indifference curves are normal to ~u)
Example

Suppose |X | = 3.

vNM utility function: U (p) = p1 u1 + p2 u2 + p3 u3

Suppose p ∼ p0 so U (p) = U (p0 ).


Example

Suppose |X | = 3.

vNM utility function: U (p) = p1 u1 + p2 u2 + p3 u3

Suppose p ∼ p0 so U (p) = U (p0 ).

By linearity,
U (αp + (1 − α)p0 ) = αU (p) + (1 − α)U (p0 ) = U (p).

So indifference curves are linear.


Example

Suppose |X | = 3.

vNM utility function: U (p) = p1 u1 + p2 u2 + p3 u3

Suppose p ∼ p0 so U (p) = U (p0 ).

By linearity,
U (αp + (1 − α)p0 ) = αU (p) + (1 − α)U (p0 ) = U (p).

So indifference curves are linear.

Suppose p00  p ∼ p0 so U (p00 ) > U (p) = U (p0 ).


Example

Suppose |X | = 3.

vNM utility function: U (p) = p1 u1 + p2 u2 + p3 u3

Suppose p ∼ p0 so U (p) = U (p0 ).

By linearity,
U (αp + (1 − α)p0 ) = αU (p) + (1 − α)U (p0 ) = U (p).

So indifference curves are linear.

Suppose p00  p ∼ p0 so U (p00 ) > U (p) = U (p0 ).

Then U (αp00 + (1 − α)p) = αU (p00 ) + (1 − α)U (p)

And U (αp00 + (1 − α)p0 ) = αU (p00 ) + (1 − α)U (p)

So indifference curves are parallel.


So, loosely, we have argued that:

Expected utility representation


⇐⇒
parallel, straight indifference curves
⇐⇒
preferences satisfy completeness, transitivity, continuity and
independence.
Which preferences have expected utility representations? I

The following is a remarkable result due to Von Neumann and


Morgenstern (1953) building on work by Bernoulli.

Theorem
A complete and transitive preference relation  on ∆(X ) satisfies
continuity and independence iff it has an expected utility
representation U : ∆(X ) → R.

Proof Outline Showing that if U (p) = p · ~u represents , then 


must satisfy continuity and independence is (relatively) easy
Showing the other direction is a bit harder. . .
Which preferences have expected utility representations? II

Roughly:
1 Let p ∈ ∆(X ) be the most preferred lottery and p ∈ ∆(X ) be
the least preferred lottery. For every p ∈ ∆(X ) find λp ∈ [0, 1]
such that
p ∼ λp p̄ + (1 − λp )p
This λp . . .
exists by continuity
is unique by independence
2 Let U (p) = λp
3 Show that U (·) is an expected utility representation of 
U (·) represents 
U (·) is linear, and therefore is a vNM utility function
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Money lotteries

We seek to measure the attitude towards risk embedded in


preferences 

We will now consider monetary payoffs


X ⊆ R (note we give up assumption that prize set is finite)
∆(X ) is now a bit more complicated
Money lotteries

We seek to measure the attitude towards risk embedded in


preferences 

We will now consider monetary payoffs


X ⊆ R (note we give up assumption that prize set is finite)
∆(X ) is now a bit more complicated
A probability distribution with finite support can be described
with a pmf; set of distributions is the simplex
A probability distribution over infinite (ordered) support is
described by a cdf

From here on, we’ll represent a lottery by a cdf F (·), where F (x)
is the probability of receiving less than or equal to an $x payout
Example

What lottery does this CDF represent?

1
Probability get less
than this

0
$0 $100
Example

What lottery does this CDF represent?

1
Probability get less
than this

0
$0 $100

You get an amount drawn uniformly from the interval [0, 100].
Example

What lottery does this CDF represent?

1
Probability get less
than this

0
$0 $100
Example

What lottery does this CDF represent?

1
Probability get less
than this

0
$0 $100

50% chance of $0 and 50% chance of $75.


Example

What lottery does this CDF represent?

1
Probability get less
than this

0
$0 $100
Example

What lottery does this CDF represent?

1
Probability get less
than this

0
$0 $100

Approximately: 50% chance of $0 and 50% chance of $75.


Probability get less
Example

than this

0
1

$0
$100
Probability get less
Example

than this

0
1
0.8

$0
$50
$100
What is the set of lotteries?

When |X | = n < ∞, the set of all lotteries is


n X o
∆(X ) ≡ p ∈ Rn+ : pi = 1
i
What is the set of lotteries?

When |X | = n < ∞, the set of all lotteries is


n X o
∆(X ) ≡ p ∈ Rn+ : pi = 1
i

When X = R, the set of all lotteries (we will consider) is the set of
cdfs:
F is the set of all functions F : R → [0, 1] such that
F (·) is nondecreasing
limx→−∞ F (x) = 0
limx→+∞ F (x) = 1
F (·) continuous (This is stronger than we need, but will
suffice for our purposes)
Preferences over money lotteries

Our old vNM utility function (over pmfs) was


X X  
U (p) = pi ui = pi u(xi ) ≡ Ep u(x)
i i
Preferences over money lotteries

Our old vNM utility function (over pmfs) was


X X  
U (p) = pi ui = pi u(xi ) ≡ Ep u(x)
i i

The continuous analogue is a vNM utility function over cdfs:


Z
 
U (F ) = u(x) dF (x) ≡ EF u(x)
R

Where
U : F → R (“von Neumann-Morganstern utility function”)
represents preferences over lotteries
u : R → R (“Bernoulli utility function”) indexes preference
over outcomes
Risk aversion I

Definition (risk aversion)


A decision-maker is risk-averse
R iff for all lotteries F , she prefers a
certain payoff of EF (x) ≡ R x dF (x) to the lottery F .

Definition (strict risk aversion)


A decision-maker is strictly risk-averse iff for all non-degenerate
lotteries F (i.e, all lotteries for which the support of F is not a
singleton),R she strictly prefers a certain payoff of
EF (x) ≡ R x dF (x) to the lottery F .
Risk aversion II

Risk aversion says that for all F ,


  
u EF [x] ≥ EF u(x)

or equivalently
Z  Z
u x dF (x) ≥ u(x) dF (x)
R R

By Jensen’s inequality, this condition holds iff u(·) is concave


Theorem
A decision-maker is (strictly) risk-averse iff her Bernoulli utility
function is (strictly) concave.
Illustrating risk aversion

Consider a risk-averse decision-maker (i.e., one with a concave


Bernoulli utility function) evaluating a lottery F with a two-point
distribution

u(x)
 
u EF (x)
 
EF u(x)

x
EF (x)

  
u EF [x] ≥ EF u(x)
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


The certain equivalent

A risk-averse decision-maker prefers a certain payoff of EF (x) to


the lottery F   
u EF [x] ≥ EF u(x)
How many “certain” dollars is F worth? That is, what is the
certain payoff that gives the same utility as lottery F ?
The certain equivalent

A risk-averse decision-maker prefers a certain payoff of EF (x) to


the lottery F   
u EF [x] ≥ EF u(x)
How many “certain” dollars is F worth? That is, what is the
certain payoff that gives the same utility as lottery F ?
Definition (certain equivalent)
The certain equivalent is the size of the certain payout such that a
decision-maker is indifferent between the certain payout and the
lottery F :
Z
  
u cu (F ) = EF u(x) ≡ u(x) dF (x).
R
Illustrating the certain equivalent

Consider a risk-averse decision-maker (i.e., one with a concave


Bernoulli utility function) evaluating a lottery F with a two-point
distribution

u(x)
 
u EF (x)
 
EF u(x)

x
cu (F )EF (x)

   
u cu (F ) = EF u(x) ≤ u EF [x]
The certain equivalent as a measure of risk aversion

For a risk-averse decision-maker


   
u cu (F ) = EF u(x) ≤ u EF [x]

so assuming increasing u(·),

cu (F ) ≤ EF [x]

The certain equivalent gives a measure of risk aversion


A consumer u is risk-averse iff cu (F ) ≤ EF [x] for all F
Consumer u is more risk-averse than consumer v iff
cu (F ) ≤ cv (F ) for all F
The Arrow-Pratt coefficient of absolute risk aversion

Definition (Arrow-Pratt coefficient of absolute risk aversion)


Given a twice differentiable Bernoulli utility function u(·),

u00 (x)
Au (x) ≡ − .
u0 (x)

Where does this come from?


Risk-aversion is related to concavity of u(·); a “more concave”
function has a smaller (more negative) second derivative
hence a larger −u00 (x)
The Arrow-Pratt coefficient of absolute risk aversion

Definition (Arrow-Pratt coefficient of absolute risk aversion)


Given a twice differentiable Bernoulli utility function u(·),

u00 (x)
Au (x) ≡ − .
u0 (x)

Where does this come from?


Risk-aversion is related to concavity of u(·); a “more concave”
function has a smaller (more negative) second derivative
hence a larger −u00 (x)
Normalization by u0 (x) takes care of the fact that au(·) + b
represents the same preferences as u(·)
We can also view it as a “probability premium”
The A-P coefficient of ARA as a probability premium

Consider a risk-averse consumer:


She prefers x for certain to a 50-50 gamble between x + ε and
x−ε
The A-P coefficient of ARA as a probability premium

Consider a risk-averse consumer:


She prefers x for certain to a 50-50 gamble between x + ε and
x−ε
If we wanted to convince her to take such a gamble, it
couldn’t be 50-50—we need to make the x + ε payout more
likely
The A-P coefficient of ARA as a probability premium

Consider a risk-averse consumer:


She prefers x for certain to a 50-50 gamble between x + ε and
x−ε
If we wanted to convince her to take such a gamble, it
couldn’t be 50-50—we need to make the x + ε payout more
likely
Consider the gamble G such that she is indifferent between G
and receiving x for certain, where
(
x+ε with probability 12 + π,
G=
x−ε with probability 12 − π
The A-P coefficient of ARA as a probability premium

Consider a risk-averse consumer:


She prefers x for certain to a 50-50 gamble between x + ε and
x−ε
If we wanted to convince her to take such a gamble, it
couldn’t be 50-50—we need to make the x + ε payout more
likely
Consider the gamble G such that she is indifferent between G
and receiving x for certain, where
(
x+ε with probability 12 + π,
G=
x−ε with probability 12 − π

It turns out that Au (x) is proportional to (π/ε) as ε → 0; i.e.,


Au (x) tells us the “premium” measured in probability that the
decision-maker demands per unit of spread ε
Our measures of risk aversion are equivalent

Theorem
The following definitions of u being “more risk-averse” than v are
equivalent:
1 Whenever u prefers F to a certain payout d, then v does as
well; i.e., for all F and d,
   
EF u(x) ≥ u(d) =⇒ EF v(x) ≥ v(d);

2 Certain equivalents cu (F ) ≤ cv (F ) for all F ;


3 u(·) is “more concave” than v(·); i.e., there exists some 
increasing concave function g(·) such that u(x) = g v(x) for
all x;
4 Arrow-Pratt coefficients of absolute risk aversion
Au (x) ≥ Av (x) for all x.
How does risk aversion change with “wealth”

Example
Suppose
 
$120 with probability 2/3
 ($110 for certain).
$60 with probability 1/3

We might then reasonably expect that


 
$220 with probability 2/3
($210 for certain).
$160 with probability 1/3
How does risk aversion change with “wealth”

Example
Suppose
 
$120 with probability 2/3
 ($110 for certain).
$60 with probability 1/3

We might then reasonably expect that


 
$220 with probability 2/3
 ($210 for certain).
$160 with probability 1/3

This is the idea of decreasing absolute risk aversion:


decision-makers are less risk-averse when they are “richer”
Decreasing absolute risk aversion

Definition (decreasing absolute risk aversion)


The Bernoulli utility function u(·) has decreasing absolute risk
aversion iff Au (·) is a decreasing function of x.
Decreasing absolute risk aversion

Definition (decreasing absolute risk aversion)


The Bernoulli utility function u(·) has decreasing absolute risk
aversion iff Au (·) is a decreasing function of x.

Definition (increasing absolute risk aversion)


The Bernoulli utility function u(·) has increasing absolute risk
aversion iff Au (·) is an increasing function of x.

Definition (constant absolute risk aversion)


The Bernoulli utility function u(·) has constant absolute risk
aversion iff Au (·) is a constant function of x.
Relative risk aversion

Definition (coefficient of relative risk aversion)


Given a twice differentiable Bernoulli utility function u(·),

u00 (x)
Ru (x) ≡ −x = xAu (x).
u0 (x)
Relative risk aversion

Definition (coefficient of relative risk aversion)


Given a twice differentiable Bernoulli utility function u(·),

u00 (x)
Ru (x) ≡ −x = xAu (x).
u0 (x)

We can define decreasing/increasing/constant relative risk aversion


as above, but using Ru (·) instead of Au (·)
DARA means that if I take a $10 gamble when poor, I will
take a $10 gamble when rich
DRRA means that if I gamble 10% of my wealth when poor, I
will gamble 10% when rich
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


When is one lottery “better than” another?

We can compare lotteries given a Bernoulli utility function


representing some preferences u

But when will two lotteries be consistently ranked under a broad


set of preferences? e.g.,
1 All nondecreasing u(·)
2 All nondecreasing, concave (i.e., risk averse) u(·)
Comparing lotteries: examples I

Example
$95 for certain vs. $105 for certain.

Example
 
$90 with probability 1/2
vs. $95 for certain.
$110 with probability 1/2
Comparing lotteries: examples II

Example
 
$90 with probability 1/2
vs. $105 for certain.
$110 with probability 1/2

Example
 
$90 with probability 1/2
vs. $110 for certain.
$110 with probability 1/2
Comparing lotteries: examples III

Example

   
$90 with probability 1/2 $80 with probability 1/2
vs. .
$110 with probability 1/2 $120 with probability 1/2
First-order stochastic dominance

Definition (first-order stochastic dominance)


Distribution G first-order stochastic dominates distribution F iff
lottery G is preferred to F under every nondecreasing Bernoulli
utility function u(·).

That is, for every nondecreasing u : R → R, the following


(equivalent) statements hold:

G u F,
   
EG u(x) ≥ EF u(x) ,
Z Z
u(x) dG(x) ≥ u(x) dF (x).
R R
Characterizing first-order stochastic dominant cdfs

Theorem
Distribution G first-order stochastic dominates distribution F iff
G(x) ≤ F (x) for all x.

That is, lottery G is more likely than F to pay at least x for any
threshold x
Probability get less
Example

than this

0
1

$0
$100
Probability get less
Example

than this

0
1
0.8

0.1
$0
$50
$100
Back to our examples. . .

Example

Lottery
x $95 $105 $110 $80 or $120 $90 or $110
< 80 F (x) = 0 0 0 0 0
[80, 90) 0 0 0 1/2 0
[90, 95) 0 0 0 1/2 1/2

[95, 105) 1 0 0 1/2 1/2

[105, 110) 1 1 0 1/2 1/2

[110, 120) 1 1 1 1/2 1


≥ 120 1 1 1 1 1
Back to our examples. . .

Example

Lottery
x $95 $105 $110 $80 or $120 $90 or $110
< 80 F (x) = 0 0 0 0 0
[80, 90) 0 0 0 1/2 0
[90, 95) 0 0 0 1/2 1/2

[95, 105) 1 0 0 1/2 1/2

[105, 110) 1 1 0 1/2 1/2

[110, 120) 1 1 1 1/2 1


≥ 120 1 1 1 1 1

($110) FOSD ($105) FOSD ($95)


($110) FOSD ($90 or $110)
Every other combination is ambiguous in terms of FOSD
Characterizing FOSD with upward shifts

Start with a lottery F and construct compound lottery G


First resolve F
Then if the resolution of F is some x, hold a second lottery
that could potentially increase (but can’t decrease) x

x f (·)

4 1
F (·)

3 1/2

2 1/2

1 1/2

0 0 x
1 2 3 4
Characterizing FOSD with upward shifts

Start with a lottery F and construct compound lottery G


First resolve F
Then if the resolution of F is some x, hold a second lottery
that could potentially increase (but can’t decrease) x

x f (·) g(·)

4 1/4 1
F (·)
G(·)
3 1/2 1/4

2 1/2

1 1/2 1/2

0 0 x
1 2 3 4
Characterizing FOSD with upward shifts

Start with a lottery F and construct compound lottery G


First resolve F
Then if the resolution of F is some x, hold a second lottery
that could potentially increase (but can’t decrease) x

x f (·) g(·)

4 1/4 1
F (·)
G(·)
3 1/2 1/4

2 1/2

1 1/2 1/2

0 0 x
1 2 3 4

G FOSD F iff we can construct G from F using upward shifts


Which of these lotteries can be ranked?

Payoff L1 L2 L3 L4
0 1/2 0 1/4 1/4
1 0 1/4 0 0
2 1/4 1/2 0 1/2
3 1/4 1/4 0 0
4 0 0 3/4 1/4
Which of these lotteries can be ranked?

Payoff L1 L2 L3 L4
0 1/2 0 1/4 1/4
1 0 1/4 0 0
2 1/4 1/2 0 1/2
3 1/4 1/4 0 0
4 0 0 3/4 1/4

L3 FOSD L4 FOSD L1 and L2 FOSD L1.


Second-order stochastic dominance

FOSD said a lottery was preferred by all nondecreasing u(·). . .

Consider whether a lottery is preferred by all risk-averse u(·)

Definition (second-order stochastic dominance)


Consider two lotteries with which pay out according to the
distributions F and G.
Distribution G second-order stochastic dominates distribution F iff
lottery G is preferred to F under every concave, nondecreasing
Bernoulli utility function u(·).
That is, for every concave, nondecreasing u : R → R,
   
EG u(x) ≥ EF u(x) .
Characterizing second-order stochastic dominant cdfs

Theorem
Distribution G second-order stochastic dominates distribution F iff
Z x Z x
G(t) dt ≤ F (t) dt for all x.
−∞ −∞
Example: Which would you prefer

1
Probability get less
than this

0
$0 $100
Example: Which would you prefer

1
Probability get less

0.8
than this

0.2
0
$0 $35 $50 $100
Example: Which would you prefer

1
Probability get less

0.8
than this

0.2
0
$0 $20 $65 $100
Probability get less
than this

0
1

$0
Example: SOSD

$100
Probability get less
than this

0
1

$0
Example: SOSD

x
$100
Probability get less
than this

0
1

$0
Example: SOSD

x
$100
Probability get less
than this

0
1

$0
Example: SOSD

x
$100
Probability get less
than this

0
1

$0
Example: SOSD

x
$100
Probability get less
than this

0
1

$0
Example: SOSD

$100
Characterizing SOSD with mean-preserving spreads I

We can construct a lottery F from G using mean-preserving


spreads if there exists a way of reaching lottery F using the
following procedure:
1. Set lottery H = G
2. Take a possible outcome of H and change it so that instead
of receiving that outcome an additional lottery is held that
has zero-mean.
3. Update lottery H to this new lottery.
4. If H = F stop. Otherwise return to step 2.
Characterizing SOSD with mean-preserving spreads II

x g(·)

4 1
G(·)

3 1/2

2 1/2

1 1/2

0 0 x
1 2 3 4
Characterizing SOSD with mean-preserving spreads III

x g(·) f (·)

4 1/4 1
G(·)
F (·)
3 1/2

2 1/4 1/2

1 1/2 1/2

0 0 x
1 2 3 4

Suppose two lotteries F and G have the same mean. Then G


SOSD F iff we can construct F from G using mean-preserving
spreads
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Demand for insurance: Setup I

Strictly risk-averse agent with wealth w

Risk of loss L with probability p

Insurance available for cost qa pays a in event of loss (agent


chooses a)

She solves
Consumer Problem

max pu[w − qa − L + a] + (1 − p)u[w − qa]


a
Demand for insurance: Setup II

The demand for insurance problem

max pu[w − qa − L + a] + (1 − p)u[w − qa]


a | {z }
≡U (a)

Note strict concavity of u(·) gives strict concavity of U (·):

U 0 (a) = (1 − q)pu0 [w − qa − L + a] − q(1 − p)u0 [w − qa]


U 00 (a) = (1 − q)2 pu00 [w − qa − L + a] + q 2 (1 − p)u00 [w − qa] < 0

Thus FOC is necessary and sufficient:

p(1 − q)u0 [w − qa∗ − L + a∗ ] = q(1 − p)u0 [w − qa∗ ]


Actuarially fair insurance

What if insurance is actuarially fair?


That is, insurer makes zero-profit: q = p
FOC becomes
 0
−
p(1

q)u [w − qa∗ − L + a∗ ] =  −
q(1  0
p)u [w − qa∗ ]
w − qa∗ − L + a∗ = w − qa∗
a∗ = L

Agent fully insures against risk of loss


Non-actuarially fair insurance

What if insurance is not actuarially fair?


Suppose cost of insurance is above expected loss: q > p
FOC is
u0 [w − qa∗ − L + a∗ ] q(1 − p)
0 ∗
=
u [w − qa ] p(1 − q)
>1
u0 [w − qa∗ − L + a∗ ] > u0 [w − qa∗ ]
w − qa∗ − L + a∗ < w − qa∗
a∗ < L

Agent under-insures against risk of loss; it’s costly to transfer


wealth to the loss state, so she transfers less
Understanding the inequality reversal in the imperfect
insurance problem

Increasing concave utility


𝑢(𝑥)

𝑥
Understanding the inequality reversal in the imperfect
insurance problem

If 𝑢(𝑥 ′ ) > 𝑢(𝑥 ′′ ) then 𝑥 ′ > 𝑥 ′′


𝑢(𝑥)

𝑥
Understanding the inequality reversal in the imperfect
insurance problem

Concavity implies decreasing


marginal utility
𝑢′ (𝑥)

𝑥
Understanding the inequality reversal in the imperfect
insurance problem

If 𝑢′ (𝑥 ′ ) > 𝑢′ (𝑥 ′′ ) then 𝑥 ′ < 𝑥 ′′


𝑢′ (𝑥)

𝑥
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Utilitarianism and interpersonal comparisons

Bentham proposed that society should maximize the sum of


individual utilities: X
Ui
i

But this implies interpersonal comparisons.

I could represent j’s preferences with a utility function


Vj = 106 Uj

I’d get different prescriptions if I instead maximized:


X
Ui + Vj ,
j6=i
A weaker criterion

Bergson (38), as a Harvard undergrad, proposed instead


maximizing a function:

W (U1 , . . . , Un ),

under the assumption that W is increasing in all its


arguments.

Does not require interpersonal comparisons.

Embeds a notion of Pareto optimality.

When can social preferences be represented in this way?


A weaker criterion

Bergson (38), as a Harvard undergrad, proposed instead


maximizing a function:

W (U1 , . . . , Un ),

under the assumption that W is increasing in all its


arguments.

Does not require interpersonal comparisons.

Embeds a notion of Pareto optimality.

When can social preferences be represented in this way?


So requires transitivity and completeness of social preferences
Transitivity is a strong assumption (e.g. Condorcet’s Paradox).
An aside: Condorcet’s Paradox

Consider the following social preferences among three


alternatives:
1/3rd of people have preferences A  B  C.
1/3rd of people have preferences C  A  B.
1/3rd of people have preferences B  C  A.

Suppose we say society prefers D to E (D S E) if more


people prefer D than E.

Then we have A S B, B S C and C S A.

Arrow’s impossibility theorem formalizes the difficulty in


generating complete and transitive social preferences.
Harsanyi’s extension

Harsanyi (1955) proposed refining Bergson’s proposal to


include uncertainty.

Showed that if all agents and society satisfy the vN-M axioms,
then society should maximize:
X
θi Ui ,
i

for individual vN-M utilities Ui and some weights θi .

As before, this representation is cardinal but does not make


interpersonal comparisons.

Can rescale the θi terms.

We are back to a (weighted) utilitarian welfare function!


Harsanyi’s argument

Consider decision making behind the veil of ignorance.

A person imagines themselves being allocated uniformly at


random to any position in society.

That person associates each position in society with some


utility vi .

Now use expected utility theory to maximize that individual’s


expected utility.
P
This implies the person maximizing: i vi .
Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Subjective probabilities

We have assumed the decision-maker accurately understands the


likelihood of each outcome

But in some cases subjective probabilities more reasonable?


Subjective probabilities

We have assumed the decision-maker accurately understands the


likelihood of each outcome

But in some cases subjective probabilities more reasonable?


Can probably agree on the probability distribution for a coin
toss.
Much harder to agree on an objective probability for the
existence of intelligent alien life.
Arguably most situations are like the second—for example,
what is the probability Trump will win the US presidential
election.
Choice with subjective probabilties

Suppose there are bookmakers taking bets on who will win


the election.
These bookmakers offer odds on each candidate winning the
election.
Modeling bets:
Set of outcomes X (e.g., dollar payouts)
Set of “states of the world” S
Bets (a.k.a. “acts”) are mappings from states of the world to
outcomes f : S → X
Suppose I am offered every possible bet and each time you
observe whether I take it or not.
What axioms might we expect my choices to satisfy?
Savage (1954)

Suppose that choices in this world satisfy axioms similar in spirit to


the vN-M axioms.
(Completeness, transitivity, something close to continuity, . . . )
Then Savage shows that choice must be must be as if maximizing
a vN-M utility function given:

Some probability distribution p over S


A Bernoulli utility function u : X → R
P P
An act f  g if and only if s∈S u(f (s))p(s) ≥ s∈S u(g(s))p(s).
Savage (1954)

The Savage result is remarkable and a key achievement in the


development of choice theory. Why?
Savage (1954)

The Savage result is remarkable and a key achievement in the


development of choice theory. Why?
It was not assumed that any probability distribution exists!
Savage (1954)

The Savage result is remarkable and a key achievement in the


development of choice theory. Why?
It was not assumed that any probability distribution exists!
We didn’t impose on the world the need to have a probability
in mind for intelligent alien life existing.
Instead, we just considered people making choices about what
bets to accept.
Savage (1954)

The Savage result is remarkable and a key achievement in the


development of choice theory. Why?
It was not assumed that any probability distribution exists!
We didn’t impose on the world the need to have a probability
in mind for intelligent alien life existing.
Instead, we just considered people making choices about what
bets to accept.
As long as those choices satisfy some (reasonable) criteria.
Choices are made as if people have some probability
distribution in mind.
Savage (1954)

The Savage result is remarkable and a key achievement in the


development of choice theory. Why?
It was not assumed that any probability distribution exists!
We didn’t impose on the world the need to have a probability
in mind for intelligent alien life existing.
Instead, we just considered people making choices about what
bets to accept.
As long as those choices satisfy some (reasonable) criteria.
Choices are made as if people have some probability
distribution in mind.
The existence of such a probability distribution is derived!
Justifies expected utility theory even when there is no
objective probability distribution. Any limitations?
Savage (1954)

Assumes you can list all possible states.


Savage (1954)

Assumes you can list all possible states.

Different people may have different probability distributions.


Outline

1 Choice Theory

2 Choice under Uncertainty


Uncertainty setup
Expected utility representation
Lotteries with monetary payoffs
Measuring risk aversion
Comparative Statics
Application: demand for insurance
Expected Utility and Social Choice
Subjective Probability
Behavioral criticisms

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Problems with the independence axiom

Example
Which lottery would you rather face?

$0 $48 $55
Lottery A 1% 66% 33%
Lottery B 100%
Problems with the independence axiom

Example
Which lottery would you rather face?

$0 $48 $55 Expected payout


Lottery A 1% 66% 33% $50
Lottery B 100% $48
Problems with the independence axiom

Example
Which lottery would you rather face?

$0 $48 $55 Expected payout


Lottery A 1% 66% 33% $50
Lottery B 100% $48
Problems with the independence axiom

Example
Which lottery would you rather face?

$0 $48 $55 Expected payout


Lottery A 1% 66% 33% $50
Lottery B 100% $48

Example
Which lottery would you rather face?

$0 $48 $55 Expected payout


Lottery C 67% 33% $18
Lottery D 66% 34% $16
Problems with the independence axiom

Example
Which lottery would you rather face?

$0 $48 $55 Expected payout


Lottery A 1% 66% 33% $50
Lottery B 100% $48

Example
Which lottery would you rather face?

$0 $48 $55 Expected payout


Lottery C 67% 33% $18
Lottery D 66% 34% $16
Illustrating Allais’ experiment

$0 $48 $55 Expected payout


Lottery A 1% 66% 33% $50
Lottery B 100% $48
Lottery C 67% 33% $18
Lottery D 66% 34% $16

$0

D C

B A
$48 $55
Alternative Models

Experimental violations of the independence axiom have generated


a huge literature looking for alternative models.
These models don’t impose linearity in probabilities.
Some have alternative axiomatic foundations.
Others just consider more general forms. For example:
X
U (p) = ui g(pi ),
i

for some increasing (but not necessarily linear) function g,


Problems with risk aversion

It seems like people are way too risk averse on small-stakes


gambles (Rabin, 2000).
Example
Would you bet on a fair coin toss where you lose $1000 or win
$1050?
Problems with risk aversion

It seems like people are way too risk averse on small-stakes


gambles (Rabin, 2000).
Example
Would you bet on a fair coin toss where you lose $1000 or win
$1050?

If you would always turn down such a bet (at any wealth level),
you would turn down a bet on a fair coin where you lose $20,000
or gain any amount

“Loss aversion” has been suggested as an explanation


Risk aversion implied by small gambles: Intuition

-100 5 110
Risk aversion implied by small gambles: Intuition

-100 5 110
Risk aversion implied by small gambles: Intuition

-1000 100M
Ellsberg Paradox

Two urns with white and black balls:


Urn A 49 black balls, 51 white balls.
Urn B also has 100 balls, each of which is black or white, but
the mix is unknown.

Three lotteries:
(i) $100 if a black ball is drawn from A, nothing otherwise.
(ii) $100 if a black ball is drawn from B, nothing otherwise.
(iii) $100 if a white ball is drawn from B, nothing otherwise.

Lotteries (ii) and (iii) have subjective risk. Lottery (i) has objective
risk.
Ellsberg Paradox

Two urns with white and black balls:


Urn A 49 black balls, 51 white balls.
Urn B also has 100 balls, each of which is black or white, but
the mix is unknown.

Three lotteries:
(i) $100 if a black ball is drawn from A, nothing otherwise.
(ii) $100 if a black ball is drawn from B, nothing otherwise.
(iii) $100 if a white ball is drawn from B, nothing otherwise.

Lotteries (ii) and (iii) have subjective risk. Lottery (i) has objective
risk.
By Savage, should prefer either (ii) or (iii) to (i), but most people
choose (i).
Ellsberg Paradox

Motivates a large behavioural literature.

Can model Urn B as having a number of white selected from some


distributions.

People might then behave as if their choice affects the distribution


selected.

For example, the worst possible distribution might always be


selected.

This allows people to be ambiguity averse.


Framing experiment 1

Example
The U.S. is preparing for an outbreak of an unusual disease, which
is expected to kill 600 people. Two alternative programs to
combat the disease have been proposed. Scientists predict that:
If program A is adopted, 200 people will be saved.
If program B is adopted, there is a 2/3 chance that no one will
be saved, and a 1/3 probability that 600 people will be saved.

Which program would you choose?


Framing experiment 1

Example
The U.S. is preparing for an outbreak of an unusual disease, which
is expected to kill 600 people. Two alternative programs to
combat the disease have been proposed. Scientists predict that:
If program A is adopted, 200 people will be saved. [72%]
If program B is adopted, there is a 2/3 chance that no one will
be saved, and a 1/3 probability that 600 people will be saved.
[28%]
Which program would you choose?
Framing experiment 2

Example
The U.S. is preparing for an outbreak of an unusual disease, which
is expected to kill 600 people. Two alternative programs to
combat the disease have been proposed. Scientists predict that:
If program C is adopted, 400 people will die with certainty.

If program D is adopted, there is a 2/3 probability that 600


people will die, and a 1/3 probability that no one will die.

Which program would you choose?


Framing experiment 2

Example
The U.S. is preparing for an outbreak of an unusual disease, which
is expected to kill 600 people. Two alternative programs to
combat the disease have been proposed. Scientists predict that:
If program C is adopted, 400 people will die with certainty.
[22%]
If program D is adopted, there is a 2/3 probability that 600
people will die, and a 1/3 probability that no one will die.
[78%]
Which program would you choose?

These two examples are exactly the same question stated in


different ways!
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Introduction

From the time of Adam Smith’s Wealth of Nations in


1776, one recurrent theme of economic analysis has been
the remarkable degree of coherence among the vast
numbers of individual and seemingly separate decisions
about the buying and selling of commodities. [...]
Would-be buyers ordinarily count correctly on being able
to carry out their intentions, and would-be sellers do not
ordinarily find themselves producing great amounts of
goods that they cannot sell. This experience of balance is
indeed so widespread that it raises no intellectual disquiet
among laymen; they take it so much for granted that
they are not disposed to understand the mechanism by
which it occurs.
Ken Arrow (1973)
Equilibrium

In a single market, we can think about price equalizing demand


and supply.

At an equilibrium price:
Consumers’ maximize their utility generating demand.
Producers’ maximize their profits generating supply.
Market clears so that demand equals supply.
But the story is a bit more complicated. . .

Demand for each good depends on prices of other goods


Marshallian demand ~x(~
p, w)

Supply also depends on the vector of prices: ~y (~


p)

General equilibrium prices satisfy

~y (~
p) = ~x(~
p, w),

Potentially a very complicated system of equations.


Have to clear all markets simultaneously.
Equilibrium Price and Efficiency

Fundamental question: Are equilibrium prices efficient?


Relevant notion of efficiency due to Pareto (1909) and
Bergson (1938)
Equilibrium Price and Efficiency

Fundamental question: Are equilibrium prices efficient?


Relevant notion of efficiency due to Pareto (1909) and
Bergson (1938)

Inquiry culminated in the Welfare Theorems (Arrow (1951)


and Debreu (1951)).
Equilibrium Price and Efficiency

Fundamental question: Are equilibrium prices efficient?


Relevant notion of efficiency due to Pareto (1909) and
Bergson (1938)

Inquiry culminated in the Welfare Theorems (Arrow (1951)


and Debreu (1951)).

Identifies a strong relationship between efficient outcomes and


market equilibria.
Equilibrium Price and Efficiency

Fundamental question: Are equilibrium prices efficient?


Relevant notion of efficiency due to Pareto (1909) and
Bergson (1938)

Inquiry culminated in the Welfare Theorems (Arrow (1951)


and Debreu (1951)).

Identifies a strong relationship between efficient outcomes and


market equilibria.

But, efficient outcomes could also be obtained in other ways.


Key advantage of market equilibria minimal informational
requirements—prices are a sufficient statistic for guiding choices.
General equilibrium: other key questions

Does a general equilibrium exist?

When is it Unique

When is it “Stable”

How does the economy reach general equilibrium prices?


An important simplification

Finding prices that equalize production and demand is hard.

We will ignore production: exchange economy


Finite number of agents
Finite number of goods
Predetermined amount of each commodity (no production)
Goods get traded and consumed

Results will extend to production economies—but this is beyond


the scope of this class.
Other assumptions

None of the following assumptions should surprise at this point,


but should be kept in mind when interpreting our following results:
Markets exist for all goods
Agents can freely participate in markets without cost
“Standard” consumer theory assumptions
Preferences can be represented by a utility function
Preferences are LNS/monotone/strictly monotone (as needed)
All agents are price takers
Finite number of divisible goods
Linear prices
Perfect information about goods and prices
All agents face the same prices
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Exchange economies: the Walrasian Model

Primitives of the model


L goods ` ∈ L ≡ {1, . . . , L}
I agents i ∈ I ≡ {1, . . . , I} (for the most part, we’ll set
I = 2)
Endowments ei ∈ RL + ; agents do not have monetary wealth,
but rather an endowment of goods which they can trade or
consume
Preferences represented by utility function ui : RL
+ →R

Endogenous prices p ∈ RL
+ , taken as given by each agent
Exchange economies: the Walrasian Model

Primitives of the model


L goods ` ∈ L ≡ {1, . . . , L}
I agents i ∈ I ≡ {1, . . . , I} (for the most part, we’ll set
I = 2)
Endowments ei ∈ RL + ; agents do not have monetary wealth,
but rather an endowment of goods which they can trade or
consume
Preferences represented by utility function ui : RL
+ →R

Endogenous prices p ∈ RL
+ , taken as given by each agent

Each agent i solves

max ui (xi ) such that p · xi ≤ p · ei ≡ max ui (xi )


xi ∈RL
+
xi ∈B i (p)

where B i (p) ≡ {xi ∈ RL i i


+ : p · x ≤ p · e } is the budget set for i
Walrasian equilibrium

Definition (Walrasian equilibrium)


Prices p and quantities (xi )i∈I are a Walrasian equilibrium iff
1 All agents maximize their utilities; i.e., for all i ∈ I,

xi ∈ argmax ui (x);
x∈B i (p)

2 Markets clear; i.e., for all ` ∈ L,


X X
xi` = ei` .
i∈I i∈I
A graphical example: the Edgeworth box

𝑒21 + 𝑒22

𝑒21 𝑒1

𝑒11 𝑒11 + 𝑒12


Agent 1
A graphical example: the Edgeworth box

𝑒21 + 𝑒22

𝑒22 𝑒2

𝑒12 𝑒11 + 𝑒12


Agent 2
A graphical example: the Edgeworth box
A graphical example: the Edgeworth box
A graphical example: the Edgeworth box

𝑒11 + 𝑒12
𝑒2

𝑒12
Agent 2
+ 𝑒22

𝑒22
1
2
A graphical example: the Edgeworth box
A graphical example: the Edgeworth box
A graphical example: the Edgeworth box

Agent 2
𝑒11 + 𝑒12 𝑒12

𝑒2 𝑒22

𝑒21 + 𝑒22
A graphical example: the Edgeworth box

Agent 2
𝑒11 + 𝑒12 𝑒12

𝑒2 𝑒22

𝑒21 + 𝑒22
A graphical example: the Edgeworth box

𝑒21 + 𝑒22

𝑒21 𝑒1

𝑒11 𝑒11 + 𝑒12


Agent 1
A graphical example: the Edgeworth box

Agent 2
𝑒12

𝑒21 𝑒22

𝑒11
Agent 1
Budget Constraints

p · x1 = p · e1 coincides with p · x2 = p · e2

𝑒21 + 𝑒22

𝑒21 𝑒1

𝑒11 𝑒11 + 𝑒12


Agent 1
Budget Constraints

p · x1 = p · e1 coincides with p · x2 = p · e2

𝑒21 + 𝑒22

𝑒22 𝑒2

𝑒12 𝑒11 + 𝑒12


Agent 2
Budget Constraints

p · x1 = p · e1 coincides with p · x2 = p · e2
Agent 2
𝑒11 + 𝑒12 𝑒12

𝑒2 𝑒22

𝑒21 + 𝑒22
Budget Constraints

p · x1 = p · e1 coincides with p · x2 = p · e2

Agent 2
𝑒11 + 𝑒12 𝑒12

𝑒2 𝑒22

𝑒21 + 𝑒22
Budget Constraints

p · x1 = p · e1 coincides with p · x2 = p · e2

Agent 2
𝑒12

𝑒21 𝑒22

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21 𝑒1

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21 𝑒1

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21 𝑒1

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21 𝑒1

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

𝑒21

𝑒11
Agent 1
The offer curve

The offer curve traces out Marshallian demand as prices change

Offer Curve

𝑒21

𝑒11
Agent 1
Non-equilibrium prices give total demand 6= supply

Agent 2

Agent 1
Equilibrium prices give total demand = supply

Agent 2

Agent 1
Equilibrium Uniqueness and Multiplicity

Agent 2

Agent 1
Equilibrium Uniqueness and Multiplicity

Agent 2

Agent 1
Equilibrium Uniqueness and Multiplicity

Agent 2

Agent 1
Some Clarifications

1 Offer curves can include infeasible allocations


They are defined just by the optimal bundle of goods for a
given agent as prices change. There is no mention of whether
that bundle is feasible or not.
2 Why are intersections of offer curves in the Edgeworth box
Walrasian equilibria whenever they are not at the endowment
point?
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Offer curves with infeasible allocations

𝑒21

𝑒11
Agent 1
Some Clarifications From Last Time

1 Offer curves can include infeasible allocations


They are defined just by the optimal bundle of goods for a
given agent as prices change. There is no mention of whether
that bundle is feasible or not.
2 Why are intersections of offer curves in the Edgeworth
box Walrasian equilibria whenever they are not at the
endowment point?
Intersections of offer curves and equilibria

𝑒21

𝑒11
Agent 1
Intersections of offer curves and equilibria

𝑒21

𝑒11
Agent 1
Intersections of offer curves and equilibria

𝑒21

𝑒11
Agent 1
Intersections of offer curves and equilibria

𝑒12

𝑒22

Agent 1
Intersections of offer curves and equilibria

𝑒12

𝑒22

Agent 1
Intersections of offer curves and equilibria

𝑒12

𝑒22

Agent 1
Intersections of offer curves and equilibria

𝑒21

𝑒11
Agent 1
Intersections of offer curves and equilibria

Agent 2

Agent 1
Intersections of offer curves and equilibria

Agent 2

Agent 1
Intersections of offer curves and equilibria

Agent 2

Agent 1
Intersections of offer curves and equilibria

Agent 2

Agent 1
Intersections of offer curves and equilibria

Agent 2

Agent 1
Intersections of offer curves and equilibria

Agent 2

Agent 1
Pareto optimality

Definition (feasible allocation)


Allocations (xi )i∈I ∈ RI·L
+ are feasible iff for all ` ∈ L,
X X
xi` ≤ ei` .
i∈I i∈I
Pareto optimality

Definition (feasible allocation)


Allocations (xi )i∈I ∈ RI·L
+ are feasible iff for all ` ∈ L,
X X
xi` ≤ ei` .
i∈I i∈I

Definition (Pareto optimality)


Allocations x ≡ (xi )i∈I are Pareto optimal iff
1 x is feasible, and
2 There is no other feasible allocation x̂ such that
ui (x̂i ) ≥ ui (xi ) for all i ∈ I with strict inequality for some i.
Pareto optimality: General case

More generally, Pareto optimality can be defined over outcomes X


for any set of agents I given notions of
1 Feasibility: a mapping X → {infeasible, feasible}
2 Individual preferences: rational preferences i over X for each
i∈I

Definition (Pareto optimality)


Outcome x ∈ X is Pareto optimal iff
1 x is feasible, and
2 There is no other feasible outcome x̂ ∈ X such that x̂ i x
for all i ∈ I with x̂ i x for some i.

This is a very weak notion of optimality, requiring only that there


is nothing “left on the table”
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Agent 2

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Agent 2

Pareto
Improvements

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Agent 2

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Agent 2

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Agent 2

No Pareto
Improvements
𝑥

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Agent 2

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Pareto Set Agent 2

Agent 1
Pareto optimality in the Edgeworth box

If the indifference curves passing through x are not tangent, it is


not Pareto optimal

Contract Curve Agent 2

Agent 1
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Relating Walrasian equilibrium and Pareto optimality

Note Walrasian Equilibria and Pareto Optima are very different


concepts
Pareto optimality

1 Allocations
given total endowments and individual preferences.

Walrasian equilibrium

1 Allocations
2 Prices
given individual endowments and preferences.
Walrasian equilibrium allocations are Pareto optimal

Agent 2

Agent 1
Walrasian equilibrium allocations are Pareto optimal

Agent 2

Agent 1
Walrasian equilibrium allocations are Pareto optimal

Agent 2

Agent 1
Walrasian equilibrium allocations are Pareto optimal

Agent 2

Agent 1
Walrasian equilibrium allocations are Pareto optimal

Agent 2

Agent 1
The First Welfare Theorem: WE are PO

Theorem (First Welfare Theorem)


Suppose ui (·) is increasing (i.e., ui (xi0 ) > ui (xi ) for any xi0  xi )
for all i ∈ I.
If p and (xi )i∈I are a Walrasian equilibrium, then the allocations
(xi )i∈I are Pareto optimal.

Proof. Suppose in contradiction that x̂ Pareto dominates x; i.e.,


1 x̂ is feasible,
2 ui (x̂i ) ≥ ui (xi ) for all i ∈ I,
0 0 0
3 ui (x̂i ) > ui (xi ) for some i0 ∈ I.
Proof (continued).

By revealed preference and Walras’ law, p · x̂i ≥ p · xi for all i, and


0 0
p · x̂i > p · xi .
Proof (continued).

By revealed preference and Walras’ law, p · x̂i ≥ p · xi for all i, and


0 0
p · x̂i > p · xi . Thus
X X
p · x̂i > p · xi
i∈I i∈I
XX XX
p` x̂i` > p` xi` .
`∈L i∈I `∈L i∈I
Proof (continued).

By revealed preference and Walras’ law, p · x̂i ≥ p · xi for all i, and


0 0
p · x̂i > p · xi . Thus
X X
p · x̂i > p · xi
i∈I i∈I
XX XX
p` x̂i` > p` xi` .
`∈L i∈I `∈L i∈I

So for some `e it must be that


X X X
x̂i`˜ > xi`˜ = ei`˜,
i∈I i∈I i∈I
Proof (continued).

By revealed preference and Walras’ law, p · x̂i ≥ p · xi for all i, and


0 0
p · x̂i > p · xi . Thus
X X
p · x̂i > p · xi
i∈I i∈I
XX XX
p` x̂i` > p` xi` .
`∈L i∈I `∈L i∈I

So for some `e it must be that


X X X
x̂i`˜ > xi`˜ = ei`˜,
i∈I i∈I i∈I

so x̂ cannot be feasible.
The Second Welfare Theorem: PO endowments are WE

Theorem (Second Welfare Theorem)


Suppose for all i ∈ I,
1 ui (·) is continuous;
2 ui (·) is increasing; i.e., ui (xi0 ) > ui (xi ) for any xi0  xi ;
3 ui (·) is concave; and
4 ei  0; i.e., every agent has at least a little bit of every good.

If (ei )i∈I are Pareto optimal, then there exist prices p ∈ Rl+ such
that p and (ei )i∈I are a Walrasian equilibrium.
Pareto optimal allocations can be supported as a
Walrasian equilibrium

Pareto Set Agent 2

Agent 1
Pareto optimal allocations can be supported as a
Walrasian equilibrium

Agent 2

Agent 1
Pareto optimal allocations can be supported as a
Walrasian equilibrium

Agent 2

Agent 1
Pareto optimal allocations can be supported as a
Walrasian equilibrium

Agent 2

Agent 1
Pareto optimal allocations can be supported as a
Walrasian equilibrium

Agent 2

Agent 1
Pareto optimal allocations can be supported as a
Walrasian equilibrium

Agent 2

Agent 1
The welfare theorems

Theorem (First Welfare Theorem)


Suppose ui (·) is increasing for all i ∈ I.
If p and (xi )i∈I are a Walrasian equilibrium, then the allocations
(xi )i∈I are Pareto optimal.

Theorem (Second Welfare Theorem)


Suppose ui (·) is continuous, increasing, and concave for all i ∈ I.
Further suppose ei  0 for all i ∈ I.
If (ei )i∈I are Pareto optimal, then there exist prices p ∈ Rl+ such
that p and (ei )i∈I are a Walrasian equilibrium.
Thoughts on the second welfare theorem

The assumptions behind the SWT are much stronger than the
FWT—in particular the requirement of convex preferences

ei  0 is required to ensure that each agent has a positive


endowment of some good with a non-zero price—that is,
everyone has non-zero wealth

The SWT is often attributed more importance than it


deserves;
Thoughts on the second welfare theorem

It is tempting to say that FWT shows price equilibria are


sufficient for efficient equilibria.
And the SWT shows price equilibria are necessary for efficient
equilibria.
Thoughts on the second welfare theorem

It is tempting to say that FWT shows price equilibria are


sufficient for efficient equilibria.
And the SWT shows price equilibria are necessary for efficient
equilibria.
But this is wrong.
SWT only shows that we can obtain a given efficient
outcomes via markets.
But if we know an outcome is efficient, why not implement it
directly?
Thoughts on the second welfare theorem

The following claim about the SWT is also commonly made:


Suppose there is a Pareto optimal allocation we want to
implement.
Then this can be done by making lump sum transfers to
redistribute wealth and then allowing the market to clear.
So, redistributional and efficiency concerns can be separated
and redistribution just comes down to lump sum transfers.
Thoughts on the second welfare theorem

But,
in practice can this really work?
Suppose we have an equilibrium in which we there is too
much inequality and we want to help the poor.
So we redistribute wealth to the poor.
But, if people’s consumption level results at least partially
from choices they make such a transfer is not lump sum.
It will distort incentives.
Without perfect information about people’s attributes and
preferences, it is very hard to find ways to make lump sum
transfers.
And people will not want to truthfully reveal this information.
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Assumptions enabling a FOC approach

Suppose for all i ∈ I,


1 ui (·) is continuous;
2 ui (·) is strictly increasing; i.e., ui (xi0 ) > ui (xi ) for any
xi0 > xi ;
3 ui (·) is concave;
4 ui (·) is differentiable; and
5 ei  0; i.e., every agent has at least a little bit of every good.
The Pareto Possibility Set

Definition (The Pareto Possibility Set)


The vectors of utility that can be achieved by feasible
consumptions:
( )
X X
1 1 n n i L i i
U = (u (x ), . . . , u (x )) : x ∈ R+ for all i, x ≤ e
i i

We will normalize the utility of consuming no goods to 0:

ui (0) = 0 for all i


The Pareto Possibility Set: An example

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Finding Pareto optimal allocations

Consider the following algorithm for finding a Pareto optimal


allocation:
1 Decide how much utility to give to each agent i ∈ {2, . . . , I}
2 Maximize agent 1’s utility subject to our decision in step 1
Finding Pareto optimal allocations

Consider the following algorithm for finding a Pareto optimal


allocation:
1 Decide how much utility to give to each agent i ∈ {2, . . . , I}
2 Maximize agent 1’s utility subject to our decision in step 1

That is,
max u1 (x1 )
x∈RIL
+

such that
i i i
P u (xi ) ≥Pū i
for i = 2, . . . , I
i∈I x` ≤ i∈I e` for ` = 1, . . . , L
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Solving the Pareto problem

The Pareto problem


Maximize u1 (x1 ) such that

x ≥ 0;
ui (xi ) ≥ ūi for i = 2, . . . , I;
i i
P P
i∈I x` ≤ i∈I e` for ` = 1, . . . , L.

Under our assumptions, all the promise-keeping and feasibility


constraints must be binding, thus multipliers > 0:
λi multiplier on ui (xi ) ≥ ūi
µ` multiplier on i xi` ≤ i ei`
P P
Solving the Pareto problem I

I
X L X
I
1 1 i
 i i i
 X
µ` (ei` − xi` ) + γ`i xi`
 
L = u (x ) + λ u (x ) − ū +
i=2 `=1 i=1

Rearranging, and letting λ1 = 1

I
X I
X L X
X I
i i i i i
µ` (ei` − xi` ) + γ`i xi`
 
L= λ u (x ) − λ ū +
i=1 i=2 `=1 i=1

Gives (summarized) FOC

∂ui
λi ≤ µ` with equality if xi` > 0
∂xi`
Solving the Pareto problem II

∂u i
Assuming x  0, the FOC is λi ∂xi = µ` , hence
`

∂ui ∂uj
∂xik µk ∂xjk
MRSik` ≡ ∂ui
= = ∂uj
≡ MRSjk`
µ`
∂xi` ∂xj`
Maximizing a Bergson-Samuelson social welfare function

Consider a planner who simply maximizes a weighted average of


individual utilities:
X
max β i ui (xi )
x∈RIL
+ i∈I

xi` ≤ i
P P
such that i∈I i∈I e` for ` = 1, . . . , L
Graphical Illustration of Bergson-Samuelson Problem

Social Planner’s Indifference Curves

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Bergson-Samuelson Problem

Changing weights

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Bergson-Samuelson Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Bergson-Samuelson Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Bergson-Samuelson Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Bergson-Samuelson Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Maximizing a Bergson-Samuelson social welfare function

Consider a planner who simply maximizes a weighted average of


individual utilities:
X
max β i ui (xi )
x∈RIL
+ i∈I
i i
P P
such that i∈I x` ≤ i∈I e` for ` = 1, . . . , L

X L X
X I
β i ui (xi ) + ψ` (ei` − xi` ) + φi` xi`
 
L=
i∈I `=1 i=1

Gives (summarized) FOC

∂ui
βi i
≤ ψ` with equality if xi` > 0
∂x`
Pareto optimality and Bergson-Samuelson SWFs

Pareto Problem:

I
X L X
X I I
X
λi ui (xi ) + µ` (ei` − xi` ) + γ`i xi` λi ūi
 
L= −
i=1 `=1 i=1 i=2
| {z }
Doesn’t depend on x

Bergson-Samuelson Problem:
I
X L X
X I
β i ui (xi ) + ψ` (ei` − xi` ) + φi` xi`
 
L=
i=1 `=1 i=1
Pareto optimality and Bergson-Samuelson SWFs

Pareto problem

xi ≥ 0
ui (xi ) ≥ ūi
P i P i
i x` ≤ i e`
∂ui
λi ∂xi ≤ µ` with equality if
`
xi` > 0
Pareto optimality and Bergson-Samuelson SWFs

Pareto problem Bergson-Samuelson problem

xi ≥ 0 xi ≥ 0
ui (xi ) ≥ ūi
P i P i
i x` ≤ i e`
P i P i i
i x` ≤
∂u
i e` β i ∂x i ≤ ψ` with equality if
i `
∂u
λi ∂xi ≤ µ` with equality if xi` > 0
`
xi` > 0
Pareto optimality and Bergson-Samuelson SWFs

Pareto problem Bergson-Samuelson problem

xi ≥ 0 xi ≥ 0
ui (xi ) ≥ ūi
P i P i
i x` ≤ i e`
P i P i i
i x` ≤
∂u
i e` β i ∂x i ≤ ψ` with equality if
i `
∂u
λi ∂xi ≤ µ` with equality if xi` > 0
`
xi` > 0

A solution to one is a solution to the other when setting


ūi = ui (xi )
λi = β i
µ` = ψ`
The Walrasian problem

A Walrasian problem
Each individual solves maxxi ∈RL ui (xi ) subject to p · xi ≤ p · ei .
P i ≤
P + i
Markets clear: x
i∈I ` i∈I e` for all ` ∈ L.

This gives Lagrangians (for the individual problems)


L
X
Li = ui (xi ) + ν i p · (ei − xi ) + ζ`i xi`
`=1
The Walrasian problem

A Walrasian problem
Each individual solves maxxi ∈RL ui (xi ) subject to p · xi ≤ p · ei .
P i ≤
P + i
Markets clear: x
i∈I ` i∈I e` for all ` ∈ L.

This gives Lagrangians (for the individual problems)


L
X
Li = ui (xi ) + ν i p · (ei − xi ) + ζ`i xi`
`=1

and (summarized) FOCs

∂ui
≤ ν i p` with equality if xi` > 0
∂xi`
The first welfare theorem

Pareto problem

xi ≥ 0
ui (xi ) ≥ ūi
P i P i
i x` = i e`
∂ui
λi ∂xi ≤ µ` with equality if
`
xi` > 0
The first welfare theorem

Pareto problem Walrasian problem

xi ≥ 0 xi ≥ 0
ui (xi ) ≥ ūi p · xi ≤ p · ei
P i P i P i P i
i x` = i e` i x` = i e`
∂ui ∂ui
λi ∂xi ≤ µ` with equality if ∂xi`
≤ ν i p` with equality if
`
xi` > 0 xi` > 0
The first welfare theorem

Pareto problem Walrasian problem

xi ≥ 0 xi ≥ 0
ui (xi ) ≥ ūi p · xi ≤ p · ei
P i P i P i P i
i x` = i e` i x` = i e`
∂ui ∂ui
λi ∂xi ≤ µ` with equality if ∂xi`
≤ ν i p` with equality if
`
xi` > 0 xi` > 0

If (x, p) is a Walrasian equilibrium, we can get that x is Pareto


optimal by setting
ūi = ui (xi )
λi = 1/ν i
µ` = p`
The second welfare theorem

Pareto problem Walrasian problem

xi ≥ 0 xi ≥ 0
ui (xi ) ≥ ūi p · xi ≤ p · ei
P i P i P i P i
i x` = i e` i x` = i e`
∂ui ∂ui
λi ∂xi ≤ µ` with equality if ∂xi`
≤ ν i p` with equality if
`
xi` > 0 xi` > 0

If x is Pareto optimal, we can get a Walrasian equilibrium (x, p) by


setting
ei = xi
ν i = 1/λi
p` = µ`
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Walrasian equilibrium

Definition (Walrasian equilibrium)


Prices p and quantities (xi )i∈I are a Walrasian equilibrium iff
1 All agents maximizing their utilities; i.e., for all i ∈ I,

xi ∈ argmax ui (x);
x∈B i (p)

2 Markets clear; i.e., for all ` ∈ L,


X X
xi` = ei` .
i∈I i∈I
Do Walrasian equilibria exist for every economy?

Theorem
Suppose for all i ∈ I,
1 ui (·) is continuous;
2 ui (·) is increasing; i.e., ui (x0 ) > ui (x) for any x0  x;
3 ui (·) is concave; and
4 ei  0; i.e., every agent has at least a little bit of every good.

There exist prices p ∈ Rl+ and allocations (xi )i∈I such that p and
x are a Walrasian equilibrium.
Excess demand

Definition (excess demand)


The excess demand of agent i is

z i (p) ≡ xi (p, p · ei ) − ei ,

where xi (p, w) is i’s Walrasian demand correspondence.


Aggregate excess demand is
X
z(p) ≡ z i (p).
i∈I

If p ∈ RL i i

+ satisfies z(p) = 0, then p and x (p, p · e ) i∈I
are a
Walrasian equilibrium
A few notes on excess demand I

X X
z(p) ≡ xi (p, p · ei ) − ei
i∈I i∈I

Under the assumptions of our existence theorem (ui (·) is


continuous, increasing, and concave, and ei  0 for all i):

z(·) is continuous
Continuity of ui implies continuity of xi
A few notes on excess demand II

X X
z(p) ≡ xi (p, p · ei ) − ei
i∈I i∈I

z(·) is homogeneous of degree zero


xi (p, wi ) is homogeneous of degree zero
i.e. xi (λp, λwi ) = λ0 xi (p, wi ) = xi (p, wi ).
xi (p, p · ei ) is homogeneous of degree zero in p
z i (p) ≡P xi (p, p · ei ) − ei is homogeneous of degree zero
z(p) ≡ i z i (p) is homogeneous of degree zero

This implies that only relative prices matter and we can normalize
one price.
A few notes on excess demand III

X X
z(p) ≡ xi (p, p · ei ) − ei
i∈I i∈I

p · z(p) = 0 for all p (Walras’ Law for excess demand)


By Walras’ Law, p · xi (p, wi ) = wi
p · xi (p, p · ei ) = p · ei 
p · z i (p) ≡ p ·Pxi (p, p · ei ) − ei = 0
p · z(p) ≡ p · i z i (p) = 0
Suppose all but one market clear; i.e., z2 (p) = · · · = zL (p) = 0

p · z(p) = p1 z1 (p) + p2 z2 (p) + · · · + pL zL (p) = 0


| {z }
=0

by Walras’ Law; hence z1 (p) = 0 as long as p1 > 0


Thus if all but one market clear, the final market must also clear
W.E. requires a solution to z(p) = 0 I

Consider a two-good economy


Normalize p2 = 1 by homogeneity of degree zero of z(·)
As long as the good one market clears, the good two market
will as well (by Walras’ Law)

We can find a W.E. whenever z1 (p1 , 1) = 0


z1 (·, 1) is continuous
As p1 → 0, excess demand for good one must go to infinity
since preferences are increasing and ei2 > 0 for all i
As p1 → ∞, excess demand for good one must be negative
since preferences are increasing and ei1 > 0 for all i
W.E. requires a solution to z(p) = 0 II

By an intermediate value theorem, there is at least one W.E.

z1

z1 (·, 1)

0 p1
p∗1
Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)
Corner Endowments-Example II

1’s utility function


Corner Endowments-Example II

1’s marginal utility (in either good)


Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)
Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)

Agent 1 has infinite marginal utility from a good when


consuming none of it.
Agent 1 will demand a strictly positive amount of good b at
any prices such that pa /pb > 0.
Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)

Agent 1 has infinite marginal utility from a good when


consuming none of it.
Agent 1 will demand a strictly positive amount of good b at
any prices such that pa /pb > 0.
Agent 2 will demand at least 30 units of good b at any prices.
Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)

Agent 1 has infinite marginal utility from a good when


consuming none of it.
Agent 1 will demand a strictly positive amount of good b at
any prices such that pa /pb > 0.
Agent 2 will demand at least 30 units of good b at any prices.
So there is no equilibrium in which pa /pb > 0.
Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)

Agent 1 has infinite marginal utility from a good when


consuming none of it.
Agent 1 will demand a strictly positive amount of good b at
any prices such that pa /pb > 0.
Agent 2 will demand at least 30 units of good b at any prices.
So there is no equilibrium in which pa /pb > 0.
If pa /pb = 0, then agent 1 will demand an infinite amount of
good a.
Corner Endowments-Example I

Suppose there are two goods a, b and two agents 1, 2.


1/2 1/2
u1 (a1 , b1 ) = a1 + b1 u2 (a2 , b2 ) = b2
e1 = (20, 0) e2 = (0, 30)

Agent 1 has infinite marginal utility from a good when


consuming none of it.
Agent 1 will demand a strictly positive amount of good b at
any prices such that pa /pb > 0.
Agent 2 will demand at least 30 units of good b at any prices.
So there is no equilibrium in which pa /pb > 0.
If pa /pb = 0, then agent 1 will demand an infinite amount of
good a.
So there is no equilibrium.
Corner Endowments and Non-Existence

20

Agent 1 b 30
Corner Endowments and Non-Existence

20

Agent 1 b 30
Corner Endowments and Non-Existence

20

Agent 1 b 30
Corner Endowments and Non-Existence

20

Agent 1 b 30
Corner Endowments and Non-Existence

20

Agent 1 b 30
Corner Endowments and Non-Existence

20

Agent 1 b 30
Corner Endowments and Non-Existence

30 b Agent 2

20
Corner Endowments and Non-Existence

30 b Agent 2

20
Corner Endowments and Non-Existence

30 b Agent 2

20
Corner Endowments and Non-Existence

30 b Agent 2

20
Corner Endowments and Non-Existence

30 b Agent 2

20
Corner Endowments and Non-Existence

b Agent 2

Agent
1
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)

For prices pa , pb > 0, 1 optimally consumes an amount


a1 = b1 .
For prices pa , pb > 0, 2 optimally consumes an amount
a2 = (b2 )2 .
Finally, in such an equilibrium a1 + a2 = 20 and b1 + b2 = 30.
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)

For prices pa , pb > 0, 1 optimally consumes an amount


a1 = b1 .
For prices pa , pb > 0, 2 optimally consumes an amount
a2 = (b2 )2 .
Finally, in such an equilibrium a1 + a2 = 20 and b1 + b2 = 30.
There is no feasible allocation that solves this system of
equations.
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

1’s Offer Curve


20

Agent 1 b 20 30
Corner Endowments-Example II

30 b Agent 2

20
Corner Endowments-Example II

30 b Agent 2

20
Corner Endowments-Example II

30 b Agent 2

20
Corner Endowments-Example II

30 b Agent 2

20
Corner Endowments-Example II

30 b Agent 2

20
Corner Endowments-Example II

30 b Agent 2

20
Corner Endowments-Example II

b Agent 2

Agent 1
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)

What if pa /pb = 0?
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)

What if pa /pb = 0?
1 has no wealth and must consume b1 = 0.
But can then optimally choose any a1 ≥ 0.
2 optimally consumes b2 = 30 and a2 ≥ 900.
Infeasible.
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)

What if pb /pa = 0?
Corner Endowments-Example II

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 30)

What if pb /pa = 0?
2 has no wealth and must consume a2 = 0.
But can then optimally choose any b2 ≥ 0.
1 optimally consumes a1 = 20 and b1 ≥ 20.
Some such allocations are feasible!
We have a continuum of Walrasian equilibria.
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

20

Agent 1 b 20 30
Corner Endowments-Example II

b Agent 2

a
Corner Endowments-Example II

b Agent 2

Agent 1
Corner Endowments-Example II

b Agent 2

Agent 1
Corner Endowments-Example II

b Agent 2

Walrasian
Equilibria a

Agent 1
Corner Endowments-Example III

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 5)
Corner Endowments-Example III

Suppose there are two goods a, b and two agents 1, 2.

u1 (a1 , b1 ) = min(a1 , b1 ) u2 (a2 , b2 ) = min((a2 )1/2 , b2 )


e1 = (20, 0) e2 = (0, 5)

For prices pa , pb > 0, 1 optimally consumes an amount


a1 = b1 .
For prices pa , pb > 0, 2 optimally consumes an amount
a2 = (b2 )2 .
Finally, in such an equilibrium

a1 + a2 = 20 and b1 + b2 = 5.

There is a feasible allocation that solves this system of


equations.
Corner Endowments-Example III

20

Agent 1 b 5
Corner Endowments-Example III

5 b Agent 2

20
Corner Endowments-Example III

b 2

a
a

1
b
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Other properties of Walrasian equilibria

We have established that an economy satisfying certain properties,


at least one Walrasian equilibrium exists

Other questions include:


1 How many Walrasian equilibria are there?

2 How does an economy (as distinct from an economist) “find”


equilibrium?

3 Can we test the Walrasian model in the data?


Uniqueness of Walrasian equilibria: Edgeworth box

Question 1: Uniqueness
Is there a unique Walrasian equilibrium (up to price
normalization)? If not, how many Walrasian equilibria are there?

x12
x21 Agent 2

Agent 1 x11
2
Uniqueness of Walrasian equilibria: Edgeworth box

Question 1: Uniqueness
Is there a unique Walrasian equilibrium (up to price
normalization)? If not, how many Walrasian equilibria are there?

x12
x21 Agent 2

Agent 1 x11
2
Uniqueness of Walrasian equilibria: Edgeworth box

Question 1: Uniqueness
Is there a unique Walrasian equilibrium (up to price
normalization)? If not, how many Walrasian equilibria are there?

x12
x21 Agent 2

OC1

Agent 1 x11
2
Uniqueness of Walrasian equilibria: Edgeworth box

Question 1: Uniqueness
Is there a unique Walrasian equilibrium (up to price
normalization)? If not, how many Walrasian equilibria are there?

x12
x21 Agent 2

OC2

Agent 1 x11
2
Uniqueness of Walrasian equilibria: Edgeworth box

Question 1: Uniqueness
Is there a unique Walrasian equilibrium (up to price
normalization)? If not, how many Walrasian equilibria are there?

x12
x21 Agent 2

OC2
OC1

Agent 1 x11
2
Uniqueness of Walrasian equilibria I

There could be one Walrasian equilibrium

z1

z1 (·, 1)

0 p1
p∗1
Uniqueness of Walrasian equilibria II

There could be two W.E. (although this is “non-generic”)

z1

z1 (·, 1)

0 p1
p∗∗
1 p∗1
Uniqueness of Walrasian equilibria III

There could be three W.E.

z1

z1 (·, 1)

0 p1
p∗∗∗
1 p∗∗
1 p∗1
Uniqueness of Walrasian equilibria IV

There could be infinite W.E. (although again, not generically)

z1

z1 (·, 1)

0 p1
Observations on multiplicity of Walrasian equilibria

It seems (and can be rigorously shown) that:


W.E. are generally not globally unique

W.E. are locally unique (generically)

There are a finite number of W.E. (generically)

There are an odd number of W.E. (generically)


Stability of Walrasian equilibria

Question 2: Stability
Is a Walrasian equilibrium “stable,” in the sense that a reasonable
dynamic adjustment process converges to equilibrium prices and
quantities?
Stability of Walrasian equilibria

Question 2: Stability
Is a Walrasian equilibrium “stable,” in the sense that a reasonable
dynamic adjustment process converges to equilibrium prices and
quantities?

Underlying question is: How does the economy “find” prices?


Hard to say in real world where prices come from
Proposed idea: a dynamic adjustment mechanism that
converges to W.E. prices
Walrasian tatonnement

One possibility
1 “Walrasian auctioneer” suggests prices
2 Agents report demand at these prices
3 If excess demand is non-zero, return to step 1
Possible price adjustment rule:

p(t + 1) = p(t) + α(t) z p(t)
Walrasian tatonnement

One possibility
1 “Walrasian auctioneer” suggests prices
2 Agents report demand at these prices
3 If excess demand is non-zero, return to step 1
Possible price adjustment rule:

p(t + 1) = p(t) + α(t) z p(t)
Big problems:
Unrealistic description of how the economy really works
No incentives to honestly report demand
Not necessarily stable
Some equilibria are unstable.
Adjustment rule can cycle and never converge.
(Scarf, H., “Some Examples of Global Instability of the
Competitive Equilibrium,” International Economic Review,
1960.)
Possible stability of Walrasian tatonnement

z1

z1 (·, 1)

0 p1
p∗1
Possible stability of Walrasian tatonnement

z1

z1 (·, 1)

0 p1
p∗1
Possible instability of Walrasian tatonnement

z1

z1 (·, 1)

0 p1
p∗∗∗
1 p∗∗
1 p∗1
Possible instability of Walrasian tatonnement

z1

z1 (·, 1)

0 p1
p∗∗∗
1 p∗∗
1 p∗1
Revived interest

Tatonnement is a very old topic.

Auction design literature has generated new interest.

How to run multi-unit auctions like the radio spectrum


auctions.

Instead of the mythical Walrasian auctioneer, there is a real


auctioneer.

Indivisible goods, but that can be worked into the theory.

Under what conditions do we get convergence to a Walrasian


equilibrium

When are agents incentivized to report their demand


truthfully.
Quasilinear Economies

Let good 1 be a numeraire good:


ui (x) = xi1 + v i (xi−1 ),
where xi−1 = (xi2 , . . . , xiL ).
Agents have enough of good 1 that the non-negativity
constraint on it never binds. Equivalently, we can allow
negative consumption of good 1.
Typically normalize p1 = 1.
Normally interpreted as partial equilibrium models.
Want to study a subset of markets and model the rest
through the numeraire good.
Only makes sense when:
wealth effects are small.
complementarities/substitutabilties with unmodeled goods are
small
Quasilinear Economies: Demand Functions

max xi1 + v i (xi−1 ) xi1 + pi−1 · xi−1 − ei−1 ≤ ei1



subject to
xi ≥0
Quasilinear Economies: Demand Functions

max xi1 + v i (xi−1 ) xi1 + pi−1 · xi−1 − ei−1 ≤ ei1



subject to
xi ≥0

X
Li = xi1 + v i (xi−1 ) + µi (ei1 − xi1 − pi−1 · xi−1 − ei−1 ) + ζ`i xi`


`6=1
Quasilinear Economies: Demand Functions

max xi1 + v i (xi−1 ) xi1 + pi−1 · xi−1 − ei−1 ≤ ei1



subject to
xi ≥0

X
Li = xi1 + v i (xi−1 ) + µi (ei1 − xi1 − pi−1 · xi−1 − ei−1 ) + ζ`i xi`


`6=1

FOC for good 1:

1 − µi = 0
Quasilinear Economies: Demand Functions

max xi1 + v i (xi−1 ) xi1 + pi−1 · xi−1 − ei−1 ≤ ei1



subject to
xi ≥0

X
Li = xi1 + v i (xi−1 ) + µi (ei1 − xi1 − pi−1 · xi−1 − ei−1 ) + ζ`i xi`


`6=1

FOC for good 1:

1 − µi = 0
So,

xi−1 ∈ argmax v i (xi−1 ) − p−1 · xi−1 − ei−1 + ei1



xi−1 ≥0

= argmax v i (xi−1 ) − p−1 · xi−1


xi−1 ≥0
Quasilinear Economies

Get demand functions

xi−1 ∈ argmax v i (xi−1 ) − p−1 · xi−1


xi−1 ≥0

Note that this demand does not depend on the consumer’s


endowment.

Pareto optimal allocations also have a much simpler form:

Consider a feasible profile (u1 , u2 , . . . , un ) ∈ U.

Then the profile (u1 + c1 , uP


2 + c2 , . . . , un + cn ) ∈ U

for any vector c such that i ci = 0.


Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Graphical Illustration of Pareto Problem

𝑢1 (𝑥 1 )

𝑢2 (𝑥 2 )
Quasilinear Economies: Pareto Optimality and Equilibria

So the planner must maximize the sum of utilities and use transfers
of the numeraire good to move around the Pareto frontier.

In all Pareto efficient allocations consumption of non-numeraire


goods is identical!

Thus in all Walrasian equilibria consumption of the non-numeraire


goods is identical.

Moreover, it can be shown that there is a unique Walrasian


equilibrium.
Testable restrictions implied by the Walrasian model

Question 3: Testability
Does Walrasian equilibrium impose meaningful restrictions on
observable data?

We noted several properties of excess demand z(p):


Continuity
Homogeneity of degree zero
Walras’ Law (p · z(p) = 0 for all p)
Limit properties
Testable restrictions implied by the Walrasian model

Question 3: Testability
Does Walrasian equilibrium impose meaningful restrictions on
observable data?

We noted several properties of excess demand z(p):


Continuity
Homogeneity of degree zero
Walras’ Law (p · z(p) = 0 for all p)
Limit properties
Actually, this is all we get
Anything goes

Theorem (Sonnenschein-Mantel-Debreu)
Consider a continuous function f : B → RL on an open and
bounded set B ⊆ RL ++ such that
f (·) is homogeneous of degree zero, and
p · f (p) = 0 for all p ∈ B.
Then there exists an economy (goods, agents, preferences, and
endowments) with aggregate excess demand function z(·)
satisfying z(p) = f (p) for all p ∈ B.

Often interpreted as “anything goes” in terms of comparative


statics. . . actually not quite right
Restrictions on possible prices

What if we can observe how equilibrium prices changes as we


change endowments. Is everything permissable?
Theorem (Brown-Matzkin)
There exist prices and endowments (p, (ei )ni=1 ) and (p̂, (êi )ni=1 )
such that it is impossible to find strictly monotone preferences
(ui )ni=1 with the property that p is a Walrasian equilibrium price
vector for the economy (ui , ei )ni=1 and p̂ is a Walrasian equilibrium
price vector for the economy (ui , êi )ni=1 .
Graphical proof, 2 goods 2 consumers

Agent 2

Agent 1
Graphical proof, 2 goods 2 consumers

Agent 2

Some point a’ on a
𝑒 is preferred by 1 to
𝑎′
any point on b
𝑎

𝑏
Agent 1
Graphical proof, 2 goods 2 consumers

Agent 2
𝑒′

Agent 1
Graphical proof, 2 goods 2 consumers

Some point c’ on c is
preferred by 1 to
any point on d
𝑑

Agent 2
𝑒′
𝑐 𝑐′

Agent 1
Graphical proof, 2 goods 2 consumers

Agent 2
By monotonicity,
point d’ is
𝑑 𝑑′
preferred by 1 to a’
𝑎′
𝑎 Agent 2
𝑏
𝑐 𝑐′

Agent 1
Graphical proof, 2 goods 2 consumers

Agent 2
By monotonicity,
point b’ is
𝑑 𝑑′
preferred by 1 to c’
𝑎′
𝑎 Agent 2
𝑏
𝑐 𝑐′
𝑏′

Agent 1
Graphical proof, 2 goods 2 consumers

Agent 2
So:
𝑎′ ≻1 𝑏′ ≻1 𝑐′ ≻1 𝑑′ ≻1 𝑎′
𝑑 𝑑′
𝑎′
𝑎 Agent 2
𝑏
𝑐 𝑐′
𝑏′

Agent 1
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Gross substitutes and the gross substitutes property I

Recall that
Definition (Gross substitute—partial equilibrium)
Good ` is a (strict) gross substitute for good m iff x` (p, w) is
(strictly) increasing in pm .

In our G.E. framework, wealth depends on prices (w = e · p) so

Definition (Gross substitute—general equilibrium)


Good ` is a (strict) gross substitute for good m iff x` (p, e · p) is
(strictly) increasing in pm .
Gross substitutes and the gross substitutes property II

Definition (Gross substitutes property)


Marshallian demand function x(p) ≡ x(p, e · p) has the (strict)
gross substitutes property if every good is a (strict) gross
substitute for every other good.

More generally. . .
Definition (Gross substitutes property)
A function f (·) has the (strict) gross substitutes property if f` (p)
is (strictly) increasing in pm for all ` 6= m.
Gross substitutes and the gross substitutes property III

Suppose each individual’s Marshallian demand satisfies the gross


substitutes property; i.e., xi` (p) is increasing in pm for all ` 6= m

Then
Individual excess demands also satisfy it: z`i (p) ≡ xi` (p) − ei` is
increasing in pm
Aggregate excess demand also satisfies it: z` (p) ≡ i z`i (p) is
P
increasing in pm
Uniqueness of Walrasian equilibrium: 2 goods

Theorem
If aggregate excess demand z(·) satisfies the strict gross
substitutes property, then the economy has at most one Walrasian
equilibrium (up to price normalization).
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Suppose λ1 > λ2 .
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Suppose λ1 > λ2 .
Let p̃1 = λ1 p1 = p01 and p̃2 = λ1 p2 > λ2 p2 = p02 .
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Suppose λ1 > λ2 .
Let p̃1 = λ1 p1 = p01 and p̃2 = λ1 p2 > λ2 p2 = p02 .
We have z1 (p0 ) = 0. Consider moving to p̃. As p01 = p̃1 this only
requires increasing p02 to p̃2 .
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Suppose λ1 > λ2 .
Let p̃1 = λ1 p1 = p01 and p̃2 = λ1 p2 > λ2 p2 = p02 .
We have z1 (p0 ) = 0. Consider moving to p̃. As p01 = p̃1 this only
requires increasing p02 to p̃2 .
By the strict gross substitutes property, increasing p2 strictly
increases demand for p1 so z1 (p̃) > 0.
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Suppose λ1 > λ2 .
Let p̃1 = λ1 p1 = p01 and p̃2 = λ1 p2 > λ2 p2 = p02 .
We have z1 (p0 ) = 0. Consider moving to p̃. As p01 = p̃1 this only
requires increasing p02 to p̃2 .
By the strict gross substitutes property, increasing p2 strictly
increases demand for p1 so z1 (p̃) > 0.
But z1 (p̃1 , p̃2 ) = z1 (λ1 p1 , λ1 p2 ) = z1 (p1 , p2 ) = 0.
Uniqueness of Walrasian equilibrium: 2 goods

Towards contradiction, two non-collinear Walrasian equilibrium


prices p = (p1 , p2 ) and p0 = (p01 , p02 ).
Let λ1 = p01 /p1 and let λ2 = p02 /p2 . By non-collinearity λ1 6= λ2 .
Suppose λ1 > λ2 .
Let p̃1 = λ1 p1 = p01 and p̃2 = λ1 p2 > λ2 p2 = p02 .
We have z1 (p0 ) = 0. Consider moving to p̃. As p01 = p̃1 this only
requires increasing p02 to p̃2 .
By the strict gross substitutes property, increasing p2 strictly
increases demand for p1 so z1 (p̃) > 0.
But z1 (p̃1 , p̃2 ) = z1 (λ1 p1 , λ1 p2 ) = z1 (p1 , p2 ) = 0. So both p and
p0 cannot be equilibria.
Uniqueness of Walrasian equilibrium: 2 goods Basic Idea

Take one of the price vectors p.


Scale it up until all prices but one are greater than the other price
vector p0 .
Now consider the good that has the same price.
There must be excess demand for this good after all the other
prices have been increased.
By the gross substitutes property and because we were in
equilibrium before.
But then we scaled up prices, without changing relative prices, and
found that we were out of equilibrium!
Works with m goods too.
Other implications of gross substitutes

The gross substitutes property can be used to show a number of


other properties of Walrasian equilibrium; e.g.,
Walrasian tatonnement will converge to the unique equilibrium

Any change that raises the excess demand of a good will


increase its equilibrium price
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium
Introduction
Exchange economies: the Walrasian Model
The Welfare theorems
Characterizing optimality and equilibrium with first-order conditions
Existence of Walrasian equilibria
Properties of Walrasian equilibria
A useful restriction: the “gross substitutes” property
General equilibrium with production
Limitations of the Welfare Theorems: Externalities
An example of inefficiency
Public goods
Producer theory: simplifying assumptions

Standard model: firms choose production plans (technologically


feasible lists of inputs and outputs) to maximize profits

Simplifying assumptions include:


1 Firms are price takers (both input and output markets)
2 Technology is exogenously given
3 Firms maximize profits; should be true as long as
The firm is competitive
There is no uncertainty about profits
Managers are perfectly controlled by owners
Production sets

Exogenously given technology applies over n commodities (both


inputs and outputs)
Definition (production plan)
A vector y = (y1 , . . . , yn ) ∈ Rn where an output has yk > 0 and
an input has yk < 0.

Definition (production set)


Set Y ⊆ Rn of feasible production plans; generally assumed to be
non-empty and closed.
Example: 1 input 1 output

y = (y1 , y2 ), where Y ≡ {y : y1 = (−y2 )0.5 , y2 ∈ [−1, 0]}


1.0

0.8

0.6

0.4

0.2

-1.0 -0.8 -0.6 -0.4 -0.2


Example: 2 inputs 1 output

y = (y1 , y2 , y3 ), where
Y ≡ {y : y1 = (−y2 )0.5 (−y3 )0.5 , y2 ∈ [−1, 0], y3 ∈ [−1, 0]}
Properties of production sets I

Definition (shutdown)
0∈Y.

Definition (free disposal)


y ∈ Y and y 0 ≤ y imply y 0 ∈ Y .

=⇒
Properties of production sets II

y = (y1 , y2 , y3 ), where
Y ≡ {y : y1 ≤ (−y2 )0.5 (−y3 )0.5 , y2 ≤ 0, y3 ≤ 0}
Properties of production sets III

Definition (nonincreasing returns to scale)


y ∈ Y implies αy ∈ Y for all α ∈ [0, 1].

Implies shutdown

Definition (nondecreasing returns to scale)


y ∈ Y implies αy ∈ Y for all α ≥ 1.

Along with shutdown, implies π(p) = 0 or π(p) = +∞ for all p

Definition (constant returns to scale)


y ∈ Y implies αy ∈ Y for all α ≥ 0; i.e., nonincreasing and
nondecreasing returns to scale.
Properties of production sets IV

Nonincreasing, Nondecreasing or Constant Returns to Scale?


Properties of production sets V

Definition (convex production set)


y, y 0 ∈ Y imply ty + (1 − t)y 0 ∈ Y for all t ∈ [0, 1].

Vaguely “nonincreasing returns to specialization”


If 0 ∈ Y , then convexity implies nonincreasing returns to scale

Strictly convex iff for t ∈ (0, 1), the convex combination is in the
interior of Y
Properties of production sets VI

(Strictly) convex?
Adding production to the Walrasian model

So far our exchange economy has treated the stock of goods


available as fixed through endowments

Now add K firms k ∈ K ≡ {1, . . . , K}


Firm k has production set Y k ⊆ RL
Will make a number of “standard” producer theory
assumptions
Also need some additional assumptions to make sure economy
is well behaved

Final primitive: what happens to firms’ profits? We typically


assume firms are owned by consumers
The Walrasian Model of the production economy I

Primitives of the model


L goods ` ∈ L ≡ {1, . . . , L}
I consumers i ∈ I ≡ {1, . . . , I}
Endowments ei ∈ RL + ; consumers do not have monetary
wealth, but rather an endowment of goods which they can
trade or consume
Preferences represented by utility function ui : RL
+ →R
K firms k ∈ K ≡ {1, . . . , K}
Production sets Y k ⊆ RL
Ownership structure (αki )k∈K,i∈I where αki is consumer i’s
share of firm k
Endogenous prices p ∈ RL
+ , taken as given by each consumer
and firm
The Walrasian Model of the production economy II

Each consumer i solves

max ui (x)
x∈B i (p)

where
n X o
B i (p) ≡ x ∈ RL i
+: p · x ≤ p · e + αki (p · y k )
k∈K

Each firm k solves


max p · y k
y k ∈Y k
The Walrasian Model of the production economy III

Definition (Aggregate production set)


The set of feasible aggregate production plans:
X X
Y ≡ Y k = {y ∈ Rn : y = y k , y k ∈ Y k for all k}.
k∈K k
Walrasian equilibrium

Definition (Walrasian equilibrium)


Prices p and quantities (xi )i∈I and (y k )k∈K are a Walrasian
equilibrium iff
1 All consumers maximize their utilities; i.e., for all i ∈ I,

xi ∈ argmax ui (x);
x∈B i (p)

2 All firms maximize their profits; i.e., for all k ∈ K,

y k ∈ argmax p · y;
y∈Y k

3 Markets clear; i.e., for all ` ∈ L,


X X X
xi` = ei` + y`k .
i∈I i∈I k∈K
Pareto optimality

Definition (feasible alocation)


Allocations (xi )i∈I ∈ RI·L k
+ and production plan (y )k∈K ∈ R
K·L
k k
are feasible iff y ∈ Y for all k ∈ K, and for all ` ∈ L,
X X X
xi` ≤ ei` + y`k .
i∈I i∈I k∈K

Definition (Pareto optimality)


Allocations (xi )i∈I and production plan (y k )k∈K are Pareto
optimal iff
1 x and y are feasible, and
2 There are no other feasible allocations x̂ and ŷ such that
ui (x̂i ) ≥ ui (xi ) for all i ∈ I with strict inequality for some i.
The First Welfare Theorem

Theorem (First Welfare Theorem)


Suppose ui (·) is increasing (i.e., ui (x0 ) > ui (x) for any x0  x) for
all i ∈ I.

If p and (xi )i∈I and (y k )k∈K are a Walrasian equilibrium, then the
allocations (xi )i∈I and (y k )k∈K are Pareto optimal.
The Second Welfare Theorem

Theorem (Second Welfare Theorem)


Suppose for all i ∈ I,
1 ui (·) is increasing; i.e., ui (x0 ) > ui (x) for any x0  x;
2 ui (·) is concave; and
3 ei  0; i.e., every agent has at least a little bit of every good.
Further suppose that production sets Y k are closed and convex for
all k ∈ K, which rules out increasing returns to scale.
Suppose (xi )i∈I and (y k )k∈K are Pareto optimal, and that xi  0
for all i ∈ I.
Then there exist prices p ∈ Rl+ , ownership shares (αki )k∈K,i∈I , and
transferred endowments (ẽi )i∈I where i ei = i ẽi such that p
P P
and (xi )i∈I and (y k )k∈K are a Walrasian equilibrium in the
economy where endowments are (ẽi )i∈I .
Do Walrasian equilibria exist for every economy?

Theorem
Suppose
ui (·) is continuous, increasing, and concave for all i ∈ I;
ei  0 for all i ∈ I;
Production sets Y k are closed and convex, and have shutdown
and free disposal for all k ∈ K;
X   X 
k
Y ∩ − Y k = {0},
k∈K k∈K

which rules out the possibility that firms can cooperate to


produce unlimited output.
Then there exists a Walrasian equilibrium.
No production pump assumption

𝑦1

𝑌𝑖

𝑦2

𝑌𝑗
No production pump assumption

𝑦1

𝑦2
No production pump assumption

𝑦1

−𝑌 𝑖

𝑦2
𝑗
−𝑌
No production pump assumption

𝑦1

−𝑌 𝑖
𝑌𝑖

𝑦2
𝑗
−𝑌
𝑌𝑗
Firms with constant returns to scale technology

Suppose a firm has CRS production technology; i.e.,


y ∈ Y =⇒ βy ∈ Y for all y and all β > 0

What can we say about its profit?


Can it be strictly positive?
Firms with constant returns to scale technology

Suppose a firm has CRS production technology; i.e.,


y ∈ Y =⇒ βy ∈ Y for all y and all β > 0

What can we say about its profit?


Can it be strictly positive? No. . . otherwise it could scale up
production arbitrarily and achieve infinite profit
Could be zero due to prices
Could be zero due to shutdown
Firms with constant returns to scale technology

Suppose a firm has CRS production technology; i.e.,


y ∈ Y =⇒ βy ∈ Y for all y and all β > 0

What can we say about its profit?


Can it be strictly positive? No. . . otherwise it could scale up
production arbitrarily and achieve infinite profit
Could be zero due to prices
Could be zero due to shutdown

So ownership structure (αki )ki doesn’t matter


General Equilibrium: Key conceptual takeaways I

Challenge: Simultaneously clearing many interrelated markets


with optimizing agents.
Concept: Walrasian equilibrium (exchange and production
economies).
Think of it just as a fixed point.
An equilibrium of the system in which everyone is happy with
their decisions.
Doesn’t prescribe how to get there.
Concept: Pareto optimality (exchange and production
economies)
First Welfare Theorem (exchange and production economies).
Second Welfare Theorem (exchange and production
economies).
General Equilibrium: Key conceptual takeaways II

Saw how the social planners problem, consumers’ problem and


Pareto problems are interlinked.
Found conditions for existence (exchange and production
economies).
Multiplicity and uniqueness under gross substitutes and in
quasilinear economies.
Tatonnenment–generally problematic, but does work under
gross substitutes.
Testability.
General Equilibrium: Key tools I

Edgeworth box.
Offer curves as a way of finding equilibria.
Pareto sets
The contract curve — Pareto optimal and individually rational.
Simplifications under Quasilinearity.
Simplifications due to Walras’ law.
Excess demand function as a way to characterize Walrasian
equilibria.
Pareto problem.
Optimization approaches to finding Walrasian equilibria under
functional form assumptions.
Simplification under gross substitutes.
Constant returns to scale and zero profits.
Implicit assumptions

Welfare theorems formalize the efficiency of markets.


Straw man.
What economically important things are assumed away?
Implicit assumptions

Welfare theorems formalize the efficiency of markets.


Straw man.
What economically important things are assumed away?
No externalities.
No missing markets.
Buyers and sellers have symmetric information.
No market power.
Externalities

Definition
An externality is present whenever the well-being of a consumer or
the production possibility of a firm are directly affected by the
actions of another agent in the economy.

The word directly is important.


Affects that occur indirectly through prices are excluded.
These are sometimes called pecuniary externalities to differentiate
them from real externalities.
Example 1: Consumption Externalities

Without any consumption externalities consumer problem is:

max ui (xi ) subject to p · xi ≤ w i .


xi ≥0
Example 1: Consumption Externalities

Without any consumption externalities consumer problem is:

max ui (xi ) subject to p · xi ≤ w i .


xi ≥0

If there are consumption externalities, i’s problem becomes

max ui (xi , x−i ) subject to p · xi ≤ w i .


xi ≥0

where x−i is the consumptions of agents other than i.


Example 2: Production Externalities

Without any production externalities firm problem is

max p · y k .
y k ∈Y k
Example 2: Production Externalities

Without any production externalities firm problem is

max p · y k .
y k ∈Y k

If there are production externalities then k’s problem becomes

max p · yk .
y k ∈Y k (y −k )

where y −k is the productions of agents other than k.


Example 3: Production Externalities

Without any production externalities consumer problem is:

max ui (xi ) subject to p · xi ≤ w i .


xi ≥0
Example 3: Production Externalities

Without any production externalities consumer problem is:

max ui (xi ) subject to p · xi ≤ w i .


xi ≥0

With production externalities, i’s problem becomes

max ui (xi , y) subject to p · xi ≤ w i .


xi ≥0

where y is the production of all firms.


Example 4: Consumption Externalities

Without any consumption externalities firm problem is

max p · y k .
y k ∈Y k
Example 4: Consumption Externalities

Without any consumption externalities firm problem is

max p · y k .
y k ∈Y k

With production externalities, i’s problem becomes

max p · y k .
y k ∈Y k (x)

where x is the consumptions of all consumers of all goods.


Consumer problem with consumption externalities

Suppose utility functions are increasing in own consumption,


differentiable in own consumption and concave in own
consumption.

max ui (xi , x−i ) subject to p · xi ≤ p · e i .


xi ≥0

This gives Lagrangians (for the individual problems)


L
X
i i i −i i i i
L = u (x , x ) + ν p · (e − x ) + ζ`i xi`
`=1
Consumer problem with consumption externalities

Suppose utility functions are increasing in own consumption,


differentiable in own consumption and concave in own
consumption.

max ui (xi , x−i ) subject to p · xi ≤ p · e i .


xi ≥0

This gives Lagrangians (for the individual problems)


L
X
i i i −i i i i
L = u (x , x ) + ν p · (e − x ) + ζ`i xi`
`=1

and (summarized) FOCs

∂ui
≤ ν i p` with equality if xi` > 0
∂xi`
Ignores the affect of xi on uj for j 6= i.
Public Goods

A special case of externalities.

Non-rivalrous, and for our purposes non-excludable.

Think of consumption of this good as a contribution to the


public good.

There are then positive externalities that take a particular


form.
Private Provision of Public goods

Move back to partial equilibrium framework for simplicity.


i
P
n consumers, one public good θ = iθ (and a numeraire good).

Utilities ui (θ) = φi (θ) − θi , where φ(·) is strictly increasing,


differentiable and concave.

max φi (θ) − θi
θi ≥0

FOC:

∂φi
−1≤0 with equality if θi 6= 0
∂θi
Private Provision of Public goods

Suppose we can order the agents in terms of their marginal


benefits so that for all aggregate provision levels θ

∂φ1 ∂φ2 ∂φn


(θ) > (θ) > · · · > (θ).
∂θ ∂θ ∂θ
Claim: Then only agent 1 contributes to the public good.
Private Provision of Public goods

Suppose we can order the agents in terms of their marginal


benefits so that for all aggregate provision levels θ

∂φ1 ∂φ2 ∂φn


(θ) > (θ) > · · · > (θ).
∂θ ∂θ ∂θ
Claim: Then only agent 1 contributes to the public good.

If agent i 6= 1 contributed to the public good:

∂φ1 ∂φi
> = 1,
∂θ ∂θ
contradicting the first order condition for 1.

Free riding!
Lindahl Equilibria

Idea: Consider each externality imposed by one agent on another.


Take each externalities and pretend it is a tradeable good.
Add a production sector where the externalities are produced as
biproducts.
For example, suppose consumption by i of good 1 generates
positive externalities for j and k.
Then add a production sector.
A unit of goods 1ii, 1ij and 1ik are produced by a unit of
good 1.
These goods replace xi1 in i, j and k’s utility functions.
Lindahl Equilibria

Find the Walrasian equilibrium in this world with no externalities.

As there are no externalities, the Walrasian equilibria are efficient.

So we have found an efficient outcome.

Conceptual importance: All externalities are just missing markets.

If we could just replace these missing markets we’d be good!


Lindahl Equilibria: Limitations

Externalities are priced bilaterally. There are no market forces to


clear these markets.
Lindahl Equilibria: Limitations

Externalities are priced bilaterally. There are no market forces to


clear these markets.
“no decentralized pricing system can serve to determine
optimally these levels of collective consumption. . . It is in
the selfish interest of each person to give false signals, to
pretend to have less interest in a given collective
consumption activity than he really has”
Samuelson, The Pure Theory of Public Expenditure (1954)
Lindahl Equilibria: Limitations

Externalities are priced bilaterally. There are no market forces to


clear these markets.
“no decentralized pricing system can serve to determine
optimally these levels of collective consumption. . . It is in
the selfish interest of each person to give false signals, to
pretend to have less interest in a given collective
consumption activity than he really has”
Samuelson, The Pure Theory of Public Expenditure (1954)

Nevertheless, there are literatures on implementing Lindahl


outcomes through bargaining and through mechanism design.
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Adverse Selection
Moral Hazard
Implicit Assumptions Underlying the Welfare Theorems

What if there is not symmetric information?

Say seller knows the quality of a good, but buyer doesn’t?

This is pre-contractual opportunism—Adverse Selection.

Say a seller of a service can chooses to provide a bad service?

This is post-contractual opportunism—Moral Hazard.

Question: Suppose someone wants a business loan. Which


type of asymmetric information should a bank worry about?
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Adverse Selection
Moral Hazard
Akerlof, 70

Akerlof, George A. “The market for “lemons”: Quality uncertainty and


the market mechanism.”
The quarterly journal of economics (1970): 488-500.
Some Examples

Buying a used car (seller has the informational advantage).

Buying insurance (buyer has the informational advantage).

TARP: Governmental program to buy “toxic assets.”

Buying a house.

Many more.
Simple Model of Health Insurance I

Can markets efficiently provide health insurance?

Welfare theorems tell us they can if certain assumptions hold.

Let’s relax one of these assumptions and see what happens.

Suppose the current health of a buyer is private information.

A buyer’s health is given by θ ∈ [0, 1].

Healthier buyers have a higher θ.

θ is uniformly distributed on [0, 1].

The buyer’s cost of treatment is 1 − θ.


Simple Model of Health Insurance II

If the buyer does not have insurance, he pays 1 − θ, otherwise


the insurance company pays.

The consumer has utility function

u(θ, y) = v(θ) + y,
where y is consumption of a numeraire good (and negative
consumption is permitted).

When i has no insurance y = −(1 − θ).

If i has insurance then y = −p, where p is the price of the


insurance.
Simple Model of Health Insurance III

We assume that v(θ) is an increasing, concave, and


differentiable function (thus that the consumer is risk averse
with respect to their health, but not the numeraire).

Insurance companies cannot differentiate between buyers, so


offer them all the same price of insurance.

Suppose the insurance market is competitive, so a price p is


charged that satisfies a zero-profit condition.

So, if all buyers purchase insurance, p = 12 .

Would all buyers purchase insurance at this price?


1
A buyer with θ > p = 2 would rather self-insure.
Simple Model of Health Insurance IV

Suppose the remaining buyers purchase insurance.


3
Then insurance companies would need to charge a price p = 4
to break even.
1
But then buyers with θ > 4 will not buy the insurance.

And so on.

The market collapses!

How can we formalize this idea?


Simple Model of Health Insurance V

When there is some consumption, we can define an


equilibrium by a p such that

Θ∗ (p) ≡ {θ : 1 − θ > p}
p = E[1 − θ|θ ∈ Θ∗ (p)]

We are looking for a price that generates zero profit


conditional on firms correctly anticipating who will buy from
them.

In the previous example, there is no such equilibrium with


Θ∗ (p) 6= ∅.
A little more realism I

But we assumed buyers know their true health.

They might be uncertain too.

What would happen if buyers observe a signal of their health


θ̂ where θ̂ = θ with probability q > 0 and with probability
1 − q it is drawn uniformly at random from [0, 1]?

Does anything change?

Consumer’s vN-M utility conditional on buying insurance:


Z 1
qv(θ̂) + (1 − q) v(s)ds − p
0
A little more realism II

Consumer’s vN-M utility conditional on not buying insurance:


Z 1
q(v(θ̂) − (1 − θ̂)) + (1 − q) v(s) − (1 − s)ds
0

Thus no insurance is preferred if and only if

1
p > q(1 − θ̂) + (1 − q) .
2
So if all consumers buy insurance, p = 1/2 but then any
consumer with signal θ̂ > 1/2 would not want to buy it.

Nothing changes! A tiny amount of private information (any


q > 0) is enough for the market to collapse.
More General Utility Functions I

Let the consumer observe her health again, but consider a


more general utility function.
Health is not a good candidate for including a numeraire.
Suppose instead we had ui (θ, y) that is increasing,
differentiable and concave in both arguments
vN-M utility conditional on buying insurance:

u(θ, −p)

Consumer’s vN-M utility conditional on not buying insurance:

u(θ, −1 + θ)
More General Utility Functions II

Insurance is then preferred if

u(θ, −p) > u(θ, −1 + θ)

As u(θ, −p) is increasing in its second argument the above


condition holds if and only if −p > −1 + θ, or equivalently

p < 1 − θ.

So nothing changes!
Combining Changes I

Consumer’s vN-M utility conditional on buying insurance:


Z 1
qu(θ̂, −p) + (1 − q) u(s, −p)ds
0

Consumer’s vN-M utility conditional on not buying insurance:


Z 1
qu(θ̂, −1 + θ̂) + (1 − q) u(s, −1 + s)ds
0
Insurance is then preferred if

Z 1
(1 − q) (u(s, −p) − u(s, −1 + s)) ds
0
 
> q u(θ̂, −1 + θ̂) − u(θ̂, −p)
Combining Changes II

If the most healthy agent takes insurance, so will all others.


Thus there is now an equilibrium in which all agents take
insurance if:

Eθ [u(θ, −1/2) − u(θ, −1 + θ)]


q
> (u(1, 0) − u(1, −1/2)) ,
1−q

Requires q fairly small (limited asymmetric information) and u


sufficiently concave (high risk aversion).
Insurmountable Problems?

The prevailing view in the 1950s was that problems of informational


asymmetry were insurmountable. Recall the Samuelson quote:
“no decentralized pricing system can serve to determine
optimally these levels of collective consumption. . . It is in
the selfish interest of each person to give false signals, to
pretend to have less interest in a given collective
consumption activity than he really has”
Samuelson, The Pure Theory of Public Expenditure (1954)

Idea: What if a planner could commit to use the information they


receive in a certain way?
Overcoming Information Asymmetries

What if we can credibly commit to how private information


will be treated?
Perhaps in a dynamic context maintaining credibility is
important enough.
Might have long lived institutions.

Can we leverage this commitment power to get efficient


outcomes?

Have people reveal all we need to know to do this?


Possible ways of creating commitment power

Contracts that will credibly be enforced by courts of law.

Politicians who will not get re-elected if they renege on


promises.
Perhaps an institution has a well defined and credible
mandate. For example:
Central banks mandated to set interest rates to target inflation.
Auction houses that commit to sell at the agree upon price.

Online reviews or word of mouth incentivise vendors to


advertise truthfully.

Can this power be used to extract enough information to


implement efficient outcomes? When? How?
General Set up

Agent: i = 1, . . . , n

Types: θi ∈ Θi

Social outcome: x ∈ X

Bernoulli utility functions: ui (θi , x).

Types are drawn from some (commonly known) distribution


according to pdf φ(θ)

And then privately observed


Planner

Asks agents to report their types.

Uses this information to choose a desirable outcome.

If the planner assumes reports will be truthful, does anything


go wrong?
Example 1: An exchange economy

L goods, n consumers and interior endowments e. So,

n X X o
L
X = (x1 , . . . , xn ) : xi ∈ R+ and x`i ≤ e`i for ` = 1, . . . , L .
i

θi is i’s preference relation over the different goods.

Θi is the set of monotone, convex preference relations over xi .

Ex-post efficient social choice rule: select a Walrasian equilibria.


Example 1: Simplified problem

Let L = n = 2, and suppose 1’s preferences are known:

Θ1 = θ1

Θ2 is the set of monotone, convex preference relations over x2

Will 2 always report report truthfully?


Example 1: Mis-reporting of Preferences

Agent 2

Agent 1
Example 1: Mis-reporting of Preferences

Agent 2

Agent 1
Example 1: Mis-reporting of Preferences

Agent 2

Agent 1
Example 1: Mis-reporting of Preferences

Agent 2

Agent 1
Example 1: Mis-reporting of Preferences

Agent 2

Agent 1
So what will work?
So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.
So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.

Commit not to make them any worse off if they tell the truth.
So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.

Commit not to make them any worse off if they tell the truth.

But maybe it is too costly to make everyone always tell us the


truth?
So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.

Commit not to make them any worse off if they tell the truth.

But maybe it is too costly to make everyone always tell us the


truth?

Can we incentivise truth telling in general?


So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.

Commit not to make them any worse off if they tell the truth.

But maybe it is too costly to make everyone always tell us the


truth?

Can we incentivise truth telling in general?

If not, then under what conditions?


So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.

Commit not to make them any worse off if they tell the truth.

But maybe it is too costly to make everyone always tell us the


truth?

Can we incentivise truth telling in general?

If not, then under what conditions?

And how?
So what will work?

If we want the truth we are going to have to incentivize


people to report truthfully.

Commit not to make them any worse off if they tell the truth.

But maybe it is too costly to make everyone always tell us the


truth?

Can we incentivise truth telling in general?

If not, then under what conditions?

And how?

And what does being correctly incentivized mean?


The right incentives

Suppose we ask people to reveal their types to us.


The right incentives

Suppose we ask people to reveal their types to us.

And commit to implement outcomes than make it in


everyone’s own self-interest to truthfully reveal their types.
The right incentives

Suppose we ask people to reveal their types to us.

And commit to implement outcomes than make it in


everyone’s own self-interest to truthfully reveal their types.

We can formulate these constraints: incentive compatibility


constraints.
The right incentives

Suppose we ask people to reveal their types to us.

And commit to implement outcomes than make it in


everyone’s own self-interest to truthfully reveal their types.

We can formulate these constraints: incentive compatibility


constraints.

We can also make sure everyone wants to participate:


individual rationality constraints
The right incentives

Suppose we ask people to reveal their types to us.

And commit to implement outcomes than make it in


everyone’s own self-interest to truthfully reveal their types.

We can formulate these constraints: incentive compatibility


constraints.

We can also make sure everyone wants to participate:


individual rationality constraints

It is relatively straightforward when these constraints don’t


depend on the actions of other agents, but more complicated
when they do. In this case we need to use game theory.
Exchange Economy

Recall the planner knows θ2 but not θ1 .


Exchange Economy

Recall the planner knows θ2 but not θ1 .

The planner wants to implement the Walrasian equilibrium, so


asks 1 to report a type.
Exchange Economy

Recall the planner knows θ2 but not θ1 .

The planner wants to implement the Walrasian equilibrium, so


asks 1 to report a type.

Suppose 1 reports θb1 , the planner then implements a


Walrasian equilibrium assuming the report is true x∗ (θ2 , θb1 ).
Exchange Economy

Recall the planner knows θ2 but not θ1 .

The planner wants to implement the Walrasian equilibrium, so


asks 1 to report a type.

Suppose 1 reports θb1 , the planner then implements a


Walrasian equilibrium assuming the report is true x∗ (θ2 , θb1 ).

Absent transfers 1 will report θb1 ∈ argmax u1 (x∗ (θ2 , θb1 ), θ1 ).


Exchange Economy

Recall the planner knows θ2 but not θ1 .

The planner wants to implement the Walrasian equilibrium, so


asks 1 to report a type.

Suppose 1 reports θb1 , the planner then implements a


Walrasian equilibrium assuming the report is true x∗ (θ2 , θb1 ).

Absent transfers 1 will report θb1 ∈ argmax u1 (x∗ (θ2 , θb1 ), θ1 ).

Thus, to make truth telling optimal for 1, we need to set


transfers such that θ1 ∈ argmax u1 (x∗ (θ2 , θb1 ), θ1 ) + t1 (θb1 ).
Mechanism Design: A preview

This is one possible solution to the planner’s problem.

Incentivize everyone to report truthfully, and then use these reports


to implement desirable outcomes.

Is this the best solution? It is efficient to always elicit truthful


reports?
Mechanism Design: A preview

This is one possible solution to the planner’s problem.

Incentivize everyone to report truthfully, and then use these reports


to implement desirable outcomes.

Is this the best solution? It is efficient to always elicit truthful


reports?

Yes! Quite generally anything you can do with any other approach,
you can also do by eliciting truthful reports.
Outline

1 Choice Theory

2 Choice under Uncertainty

3 General Equilibrium

4 Limitations of the Welfare Theorems: Asymmetric information


Adverse Selection
Moral Hazard
Some of the key early contributors

Holmstrom Milgrom Mirrlees Stiglitz Zeckhauser


Some Examples

Labor: How much effort workers exert.

Traders who are paid large bonuses.

Mangers who want to “empire build” to further their careers.

Equity holders of failing firms.

Insurance.

Many more.
Contract Theory I

Maybe writing contracts before actions are taken can limit moral
hazard difficulties?
A manager employs a worker.
Worker exerts costly effort to produce output. But, the effort
exerted by the worker is private information.
Manager wants to incentivize the worker to worker hard, but
the worker prefers to slack off.
First the manager commits to a contract for the worker.
The worker decides whether to accept the contract.
If so the worker chooses effort e ∈ R+ at cost c(e) = e2 /2.
Contract Theory II

Output is then realized, y = e + ε


ε is a normally distributed, mean 0 variance σ 2 noise term.
ε is not observed, the agent observes e but principal doesn’t.
Thus the contract offered by the manager depends only on y.
Let w(y) be this wage offer.
Firm is risk neutral with payoff:

UM = E[y − w(y)].

Worker is risk averse with payoff:

UW = E[w(y) − c(e)] − (λ/2) Var[w − c(e)].


If effort was observed I

Special case, manager observes effort


Consider the effort that maximizes surplus:

e∗ ≡ argmax E[y − c(e)] = argmax e − e2 /2 = 1


e e

Can write a contract that only pays a worker who exerts this effort
level.
e∗2
So w = 0 if e 6= e∗ and otherwise, w = 2 .

Worker problem is then:

max E[w(e) − c(e)] − (λ/2) Var[w(e) − c(e)].


e

A worker’s expected payoff is 0 for e ∈ {e∗ , 0} and negative


otherwise.
If effort was observed II

We resolve the worker’s indifference in favor of e∗ , but this is not


important. Why?
Thus the manager incentivizes the worker to maximize the surplus,
and then extracts it all.
Linear Contracts I

Suppose the manager offers a linear contract:

w(y) = α + βy
Useful to solve from the back and work forwards.
In the second stage, the worker chooses effort to solve:

max E[α + β(e + ε) − c(e)] − (λ/2) Var[α + β(e + ε)]


e∈R+

simplifies to
max α + βe − e2 /2 − (λ/2)β 2 σ 2
e∈R+

This problem is concave, so the first order conditions will


characterize the optimum.
The first order condition requires that β − e = 0
Linear Contracts II

Worker’s expected utility from accepting the contract and


optimizing effort is then:

β2
α + β 2 − β 2 /2 − (λ/2)β 2 σ 2 = α + (1 − λσ 2 )
2
Accepting the contract is optimal for worker if and only if:

1 λσ 2
α + β2( − )≥0
2 2
We refer to this as the paticipation or individual rationality
constraint for the worker.
Linear Contracts III

Now to the first stage.


Manager anticipates worker response to contract w(y) = α + βy
Selects contract to maximize her payoff condition on this.
There are many ways to solve this problem.
We could use the Lagrangian method.
But, is time consuming—must find which constraints bind.
We’ll take a different approach, using “economic intuition.”

Eye ball the expected utility of the worker and the manager:
Linear Contracts IV

UM (y) = E [y] − E [α + βy]


= (1 − β)E [y] − α
λ
UW (y) = E [w(y)] − Var [w(y)] − c(e)
2
λ 2 
= α + βE [y] − β Var [y] − c(e)
2
Where does α enters these expressions?
There’s a ‘−α’ term in the manager’expected s utility and a ‘+α’
term in the worker’s expected utility.
Moreover, UM (y) + UW (y) doesn’t depend on α.
Linear Contracts V

The manager can costlessly transfer utility between himself and the
worker!
So: set β to maximize joint surplus, subject to incentive
compatibility, then transfer it all to herself using α

 
2 λ 2 1
UM (y) + UW (y) = β − β σ +
2 2
Linear Contracts VI

So we solve the following problem:

β2
max β − (λσ 2 + 1),
β∈R 2
FOC implies that β = 1/(1 + λσ 2 ). So e = 1/(1 + λσ 2 ) < 1.
So the optimal (linear contract) places less emphasis on output
when the principals observation of effort is worse and when the
agent is more risk averse.
Now, to extract all the gains from trade make UW (y) = 0 by
setting

β2
α=− (1 − λσ 2 )
2
Linear Contracts VII

Although α is negative, we normalized the worker’s values to


rejecting the contract to 0.
So in practice we might expect to see an α > 0 but less than this
outside option.
Note that the manager can infer what effort the agent has taken.
Why not just directly compensate the worker for his effort?
This would avoid transferring risk to him.
However, the manager cannot do this.
If the manager inferred effort and rewarded this, the agent would
want to deviate.
The agent would put in low effort and claim a bad shock
happened.
Linear Contracts VIII

Remark
To create the right incentives, the principal has to punish agents
who receive bad shocks and reward agents who receive god shocks,
even though he knows these deviations are just because of the
shocks.
Linear Contracts IX

Why is the principal constrained to use linear contracts?

What if the agent performs multiple tasks with different levels


of observability and his allocation of effort to each tasks is
unobserved?

Empirically evidence is mixed. Incentives do seem to matter in


a variety of contexts, but the theory has not been especially
successful at predicting the optimal form of contracts.

You might also like