EC400 Lecture Notes
EC400 Lecture Notes
Course Outline
• Quadratic forms are useful because: (i) the simplest functions after linear ones;
(ii) conditions for optimization techniques are stated in terms of quadratic
forms; (iii) economic optimization problems have a quadratic objective func-
tion, such as risk minimization problems in finance, where riskiness is measured
by the quadratic variance of the returns from investments.
• Among the functions of one variable, the simplest functions with a unique global
extremum are the pure quadratics: y = x2 and y = −x2 . The level curve of a
general quadratic form in R2 is
and can take the form of an ellipse, a hyperbola, a pair of lines, or possibly, the
empty set.
2
• Each quadratic form can be represented as
Q(x) = xT Ax
3
Definiteness of quadratic forms
then if a > 0, ax2 is non negative and equals 0 only when x = 0. This is positive
definite, and x = 0 is a global minimizer. If a < 0, then the function is negative
definite.
• In two dimensions,
x21 + x22
• There are two intermediate cases: if the quadratic form is always non negative
but also equals 0 for non zero x0 s, is positive semidefinite, such as
(x1 + x2 )2
which can be 0 for points such that x1 = −x2 . A quadratic form which is never
positive but can be zero at points other than the origin is called negative
semidefinite.
• We apply the same terminology for the symmetric matrix A, that is, the matrix
A is positive semi definite if Q(x) = xT Ax is positive semi definite and so on.
4
• Definition: let A be an (n × n) symmetric matrix. Then A is:
(a) positive definite if xT Ax > 0 for all x 6= 0 in Rn ,
(b) positive semi definite if xT Ax ≥ 0 for all x 6= 0 in Rn ,
(c) negative definite if xT Ax < 0 for all x 6= 0 in Rn ,
(d) negative semi definite if xT Ax ≤ 0 for all x 6= 0 in Rn ,
(e) indefinite xT Ax > 0 for some x 6= 0 in Rn and xT Ax < 0 for some x 6= 0 in
Rn .
! ! !
a22 a23 a21 a23 a21 a22
a11 det − a12 det + a13 det .
a32 a33 a31 a33 a31 a32
5
• Definition: Let A be an (n × n) matrix. A (k × k) submatrix of A formed by
deleting (n − k) columns, say columns (i1 , i2 , ..., in−k ) and the same (n − k) rows
from A, (i1 , i2 , ..., in−k ) , is called a kth order principal submatrix of A. The
determinant of a (k × k) principal submatrix is called a kth order principal
minor of A.
there is one third order principal minor, which is det(A). There are three second
ordered principal minors and three first order principal minors.
– A is positive definite if and only if all its n leading principal minors are
strictly positive.
– A is negative definite if and only if all its n leading principal minors alter-
nate in sign as follows:
The kth order leading principal minor should have the same sign as (−1)k .
– A is positive semidefinite if and only if every principal minor of A is non
negative.
6
– A is negative semidefinite if and only if every principal minor of odd order
is non positive and every principal minor of even order is non negative.
• Diagonal matrices:
a1 0 0
A=
0 a2 0 .
0 0 a3
These also correspond to the simplest quadratic forms:
This quadratic form will be positive (negative) definite if and only if all the
a0i s are positive (negative). It will be positive semidefinite if and only if all the
ai ; s are non negative and negative semidefinite if and only if all the a0i s are non
positive. If there are two a0i s of opposite signs, it will be indefinite.
b2 2 b2 2
Q(x1 , x2 ) = ax21 + 2bx1 x2 + cx22 + x − x2
a 2 a
2 2
2bx1 x2 b 2 b
= a(x21 + + 2 x2 ) − x22 + cx22
a a a
b 2 (ac − b2 ) 2
= a(x1 + x2 ) + x2
a a
7
• If both coefficients, a and (ac−b2 )/a are positive, then Q will never be negative.
It will equal 0 only when x1 + ab x2 = 0 and x2 = 0 in other words, when x1 = 0
and x2 = 0. In other words, if
a b
|a| > 0 and det A = >0
b c
Similarly, Q will be negative definite if and only if both coefficient are negative,
which occurs if and only if a < 0 and ac − b2 > 0, that is, when the leading
principal minors alternative in sign. If ac − b2 < 0. then the two coefficients will
have opposite signs and Q will be indefinite.
• Examples of (2 × 2) matrixes:
!
2 3
– Consider A = . Since |A1 | = 2 and |A2 | = 5, A is positive
3 7
definite.
!
2 4
– Consider B = . Since |B1 | = 2 and |B2 | = −2, B is indefinite.
4 7
Taylor’s formulation:
f (a + h) ≈ f (a) + f 0 (a)h
8
R(h; a) is the difference between the two sides of the approximation, and by the
R(h;a)
definition of the derivative f 0 (a), we have h
→ 0 as h → 0.
where
Rk (h; a)
f (a + h) − Pk (a + h) = Rk (h; a) where lim =0
h→0 hk
• Example: we compute the first and second order Taylor polynomial of the
exponential function f (x) = ex at x = 0. All the derivatives of f at x = 0 equal
1. Then:
P1 (h) = 1 + h
h2
P2 (h) = 1 + h +
2
For h = .2, then P1 (.2) = 1.2 and P2 (.2) = 1.22 compared with the actual value
of e.2 which is 1.22140.
∂F ∂F
F (a + h) ≈ F (a) + (a)h1 + ... + (a)hn
∂x1 ∂xn
R1 (h;a)
where ||h||
→ 0 as h → 0. This is the approximation of order 1. Alternatively
9
∂F ∂F
where DFa = ∂x1
(a), ..., ∂xn
(a) .
f 00 (a) 2
For order two, the analogue for 2!
h is
1 T 2
h D Fa h,
2
∂2F ∂2F
∂ 2 x1
... ∂xn ∂x1
x=a x=a
D 2 Fa = ... ... ... .
∂2F ∂2F
∂x1 ∂xn
... ∂ 2 xn
x=a x=a
10
Lecture 2: Unconstrained optimization.
• Definition: The ball B(x, r) centred at x of radius r is the set of all vectors y
in Rn whose distance from x is less than r, that is
∂F ∂F
= 3x2 + 9y = 0; = −3y 2 + 9x = 0
∂x ∂y
11
the critical points are (0, 0) and (3, −3).
for all x ∈ C.
12
Functions of several variables
∂f (x∗ )
= 0 for i = 1, 2, ..., n.
∂xi
Can we say whether (0, 0) or (3, −3) are a local maximum or a local minimum
then? For this we have to consider the Hessian, or the matrix of the second
order partial derivatives. Note that this is a symmetric matrix since cross-
partial derivatives are equal (if the function has continuous second order partial
derivatives, Clairaut’s / Schwarz’s theorem).
But it is not true that if x∗ is a critical point, and D2 f (x∗ ) is negative (positive)
semidefinite, then x∗ is a local maximum. A counterexample is f (x) = x3 ,
which has the property that D2 f (0) is semidefinite, but x = 0 is not a maximum
or minimum.
13
• Back to the example of F (x, y) = x3 − y 3 + 9xy. Compute the Hessian:
!
6x 9
D2 F (x, y) =
9 −6y
The first order leading principle minor is 6x and the second order leading princi-
pal minor is det (D2 F (x, y)) = −36xy − 81. At (0, 0) these two minors are 0 and
−81 and hence the matrix is indefinite and this point is neither a local min or
a local max (it is a saddle point). At (3, −3) these two minors are positive and
hence it is a strict local minimum of F. Note that it is not a global minimum
(why?).
• Sketch of proof:
1
F (x∗ + h) = F (x∗ ) + DF (x∗ )h + hT D2 F (x∗ )h + R(h)
2
1
F (x∗ + h) − F (x∗ ) ≈ hT D2 F (x∗ )h
2
If D2 F (x∗ ) is negative definite, then for all small enough h 6= 0, the right hand
side is negative. Then
F (x∗ + h) < F (x∗ )
14
A real valued function g defined on a convex subset U of Rn is convex, if for all
x, y in U and for all t ∈ [0, 1] :
• A convex set:
Definition: A set U is a convex set if for all x ∈ U and y ∈ U, then for all
t ∈ [0, 1] :
tx + (1 − t)y ∈ U
• Concave and convex functions need to have convex sets as their domain. Oth-
erwise, the conditions above fail.
∂f (x) ∂f (x)
f (y) − f (x) ≤ Df (x)(y − x) = (y1 − x1 ) + ... + (yn − xn )
∂x1 ∂xn
15
• If f is a continuous and differentiable concave function on a convex set U and
if x0 ∈ U, then
Df (x0 )(y − x0 ) ≤ 0
implies f (y) ≤ f (x0 ), and if this holds for all y ∈ U, then x0 is a global
maximizer of f .
Hence also
f (y) − f (x0 ) ≤ 0.
16
The property that critical points of concave functions are global maximizers is
an important one in economic theory. For example, many economic principles,
such as marginal rate of substitution equals the price ratio, or marginal revenue
equals marginal cost are simply the first order necessary conditions of the corre-
sponding maximization problem as we will see. Ideally, as economist would like
such a rule also to be a sufficient condition guaranteeing that utility or profit
is being maximized, so it can provide a guideline for economic behaviour. This
situation does indeed occur when the objective function is concave.
17
Lecture 3: Concavity, convexity, quasi-concavity and economic
applications
• Recall:
Definition: A set U is a convex set if for all x ∈ U and y ∈ U, then for all
t ∈ [0, 1] :
tx + (1 − t)y ∈ U
• Concave and convex functions need to have convex sets as their domain.
• Let f1 , ..., fk be concave functions, each defined on the same convex subset U
of Rn . Let a1 , a2 , ..., ak be positive numbers. Then a1 f1 + a2 f2 + ... + ak fk is a
concave function on U.
(Proof: in class).
Consider the problem of maximizing profit for a firm whose production function
is y = g(x), where y denotes output and x denote the input bundle. If p denotes the
price of output and wi is the cost per unit of input i, then the firm’s profit function is
18
The first order conditions:
∂g
p = wi for i = 1, 2, ..., n
∂xi
Xaf = {x ∈ U |f (x) = a}
Ca+ = {x ∈ U |f (x) ≥ a}
is a convex set.
Thus, the level sets of the function bound convex subsets from below.
Ca− = {x ∈ U |f (x) ≤ a}
is a convex set.
Thus, the level sets of the function bound convex subsets from above.
19
Proof: Let x and y be two points in Ca+ so that f (x) ≥ a and f (y) ≥a. Then
So tx + (1 − t)y is in Ca+ and hence this set is convex. We have shown that if
f is concave, it is also quasi-concave. Try to show that every convex function
is quasi-convex.
• The property that the set above any level set of a concave function is a convex
set is a natural requirement for utility and production functions. For example,
consider an indifference curve C of the concave utility function U . Take two
bundles on this indifference curve. The set of bundles which are preferred to
them, is a convex set. In particular, the bundles that mix their contents are
in this preferred set. Then, given any two bundles, a consumer with a concave
utility function will always prefer a mixture of the bundles to any of them.
20
decreasing function.
• The region above and to the right of any of this function’s level sets is a convex
set and hence Q is quasi-concave.
21
Lecture 4: Constrained Optimization I: The Lagrangian
• We now analyze optimal allocation in the presence of scarce resources; after all,
this is what economics is all about.
and
h1 (x1 , x2 , ..., xn ) = c1 , .., hm (x1 , x2 , ..., xn ) = cm .
• The function f is called the objective function, while the g and h functions are
the constraint functions: inequality constraint (g) and equality constraints (h).
subject to
p1 x1 + p2 x2 + ... + pn xn ≤ I
x1 ≥ 0, x2 ≥ 0, ..., xn ≥ 0
22
Equality constraints:
max f (x1 , x2 )
x1 ,x2
subject to
p 1 x 1 + p 2 x2 = I
f (x1 , x2 ) = a
∂f (x1 , x2 ) ∂f (x1 , x2 )
dx1 + dx2 = 0
∂x1 ∂x2
Then:
dx2 ∂f (x1 , x2 ) ∂f (x1 , x2 )
=− /
dx1 ∂x1 ∂x2
∂f ∗ ∂f ∗
− (x )/ (x )
∂x1 ∂x2
23
The slope of the constraint at x∗ is
∂h ∗ ∂h ∗
− (x )/ (x )
∂x1 ∂x2
or:
∂f ∂f
∂x1
(x∗ ) ∂x2
(x∗ )
∂h
= ∂h
∂x1
(x∗ ) ∂x2
(x∗ )
∂f ∂f
∂x1
(x∗ ) ∂x2
(x∗ )
∂h
= ∂h
=µ
∂x1
(x∗ ) ∂x2
(x∗ )
∂f ∗ ∂h ∗
(x ) − µ (x ) = 0
∂x1 ∂x1
∂f ∗ ∂h ∗
(x ) − µ (x ) = 0
∂x2 ∂x2
∂f ∗ ∂h ∗
(x ) − µ (x ) = 0
∂x1 ∂x1
∂f ∗ ∂h ∗
(x ) − µ (x ) = 0
∂x2 ∂x2
h(x∗1 , x∗2 ) = c
24
and then find the critical point of L, by setting:
∂L
= 0
∂x1
∂L
= 0
∂x2
∂L
= 0
∂µ
An example:
max (x1 x2 )
x1 ,x2
subject to
x1 + 4x2 = 16
25
L(x1 , x2 , µ) = x1 x2 − µ(x1 + 4x2 − 16)
x2 − µ = 0
x1 − 4µ = 0
x1 + 4x2 − 16 = 0
26
Inequality constraints:
∂f ∗ ∂h ∗
(x ) − µ (x ) = 0
∂x1 ∂x1
∂f ∗ ∂h ∗
(x ) − µ (x ) = 0
∂x2 ∂x2
Or: !
∂f
∂x1
(x∗ ) ∂h
∂x1
(x∗ )
∂f
=µ
∂x2
(x∗ ) ∂h
∂x2
(x∗ )
Or:
∇f (x∗ ) = µ∇h(x∗ ).
max f (x1 , x2 )
x1 ,x2
subject to
g(x1 , x2 ) ≤ b
Graphical representation: In the graph, the solution is where the level curve of
f meets the boundary of the constraint set. This means that the constraint is
binding. There is a tangency at the solution.
• But now when we look graphically at the constraint optimization problem, even
when the constraint is binding, we would have a restriction on the Lagrange
27
multiplier. The gradients are again in line so that one is multiplier of the other:
∇f (x∗ ) = λ∇g(x∗ ).
But now the sign of λ is important: the gradients must point in the same direction
also because otherwise we can increase f and still satisfy the constraint. This means
that λ ≥ 0. This is the main difference between inequality and equality constraints.
We still form the Lagrangian:
∂L ∂f ∂g
= −λ =0
∂x1 ∂x1 ∂x1
∂L ∂f ∂g
= −λ =0
∂x2 ∂x2 ∂x2
∂L
But what about ∂λ
?
Suppose that the optimal solution is when g(x1 , x2 ) < b. At this point, the con-
straint is not binding, as the optimal solution is at the interior. The point x∗ of the
optimal solution is a local maximum (it is an unconstrained maximum). Thus:
∂f ∗ ∂f ∗
(x ) = (x ) = 0
∂x1 ∂x2
λ(g(x1 , x2 ) − b) = 0.
28
Lecture 5: Constrained Optimization II: Inequality Constraints
∂L(x∗ , λ∗ )
= 0
∂x1
∂L(x∗ , λ∗ )
= 0
∂x2
λ∗ (g(x∗1 , x∗2 ) − b) = 0
λ∗ ≥ 0
g(x∗1 , x∗2 ) ≤ b
An example:
29
The first order condition is:
x−.5 − 1 − λ = 0
x−.5 − 1 − λ = 0
λ(x − a) = 0
λ ≥ 0
x ≤ a
1
x = a, λ = √ − 1
a
When is this solution viable? We need to keep consistency so if we assume that λ > 0
then we need to insure it:
1
√ −1>0⇔a<1
a
What if λ = 0? this means that the constraint is not binding. From the first order
condition:
x−.5 − 1 = 0 ⇔ x = 1
30
Several Inequality constraints:
The generalization is easy: however, now some constraints may be binding and
some may be not binding.
An example:
We have to maximize f (x, y, z) = (xyz) subject to the constraints that x+y+z ≤ 1
and that x ≥ 0, y ≥ 0 and z ≥ 0. The Lagrangian is
xyz − λ1 (x + y + z − 1) + λ2 x + λ3 y + λ4 z
Solving the Lagrange problem will give us a set of critical points. The optimal
solution will be a subset of this. But we can already restrict this set of critical points
because it is obvious that λ2 = 0 = λ3 = λ4 . If one of these is positive, for example
λ2 > 0, then it must mean by complementary slackness, that x = 0. But then the
value of xyz is 0, and obviously we can do better than that (for example, when
x = y = z = .1).
Thus, the non-negativity conditions cannot bind. This leaves us with a problem
with one constraint, and we have to decide whether λ1 > 0 or λ1 = 0. But obviously,
the constraint must bind. If x + y + z < 1 we can increase one of the variables, satisfy
the constraint, and increase the value of the function. From the first order conditions:
xy − λ1 = 0
zy − λ1 = 0
xz − λ1 = 0
1
we then find that xy = yz = zx and hence it follows that x = y = z = 3
at the
optimal solution.
31
We have looked at: max f (x, y) subject to g(x, y) ≤ b..
We have characterized necessary conditions for a maximum. So that if x∗ is
a solution to a constrained optimization problem (it maximizes f subject to some
constraints), it is also a critical point of the Lagrangian. We find the critical points
of the Lagrangian.
• Can we then say that these are the solutions for the constrained optimization
problem? In other words:
• Can we say that these are maximizers of the Lagrangian, and if these are max-
imizers of the Lagrangian, are these also maximizers of f (subject to the con-
straint)?
To determine the answer, let (x0 , y 0 , λ) satisfy all necessary conditions for a max-
imum. It is clear that if x0 , y 0 is a maximizer of the Lagrangian, it also maximizes
f.
To see this note that λ[g(x0 , y 0 )−b] = 0. Thus, f (x0 , y 0 ) = f (x0 , y 0 )−λ(g(x0 , y 0 )−b).
By λ ≥ 0 and g(x, y) ≤ b for all other (x, y), then f (x, y) − λ(g(x, y) − b) ≥ f (x, y).
Since x0 , y 0 maximizes the Lagrangian, then for all other x, y :
32
• If f is a concave function defined on a convex subset X in Rn , x0 is a point
in the interior in which Df (x0 ) = 0, then x0 maximizes f (x) in X, that is,
f (x) ≤ f (x0 ) for all x.
33
The Kuhn-Tucker Theorem:
Consider the problem of maximizing f (x) subject to the constraint that g(x) ≤ b.
Assume that f and g are differentiable, f is concave, g is convex, and that the
constraint qualification holds. Then x∗ solves this problem if and only if there is a
scalar λ such that
∂L(x∗ , λ) ∂ ∂
= f (x∗ ) − λ g(x∗ ) = 0 for all i
∂xi ∂xi ∂xi
λ ≥ 0
g(x∗ ) ≤ b
λ[b − g(x∗ )] = 0
Mechanically (that is, without thinking...), one can solve constrained optimization
problems in the following way:
• Suppose that there exist λ∗ such that the first order conditions are satisfied,
that is:
∂L(x∗ , λ∗ )
= 0 for all i
∂xi
λ∗ ≥ 0
λ∗i (g(xi ) − b) = 0
• Assume that g1 to ge are binding and that ge+1 to gm are not binding. Write
(g1 , .., ge ) as gE . Assume also that the Hessian of L with respect to x at x∗ , λ∗
is negative definite on the linear constraint set {v : DgE (x∗ )v = 0}, that is:
34
• Then x∗ is a strict local constrained max of f on the constraint set.
If the last n − e leading principal minors of Q alternate in sign with the sign
of the determinant of the largest matrix the same as the sign of (−1)n , then
sufficient second order conditions hold for a candidate point to be a solution of
a constrained maximization problem.
35
Lecture 6: Constrained Optimization III: Maximum value functions
Profit functions and indirect utility functions are example of maximum value func-
tions, whereas cost functions and expenditure functions are minimum value functions.
If x(b) solves the problem of maximizing f (x) subject to g(x) ≤ b, the maximum
value function is v(b) = f (x(b)).
where b∗ = (b∗1 , ..., b∗k ). Let x∗1 (b∗ ), ..., x∗n (b∗ ) denote the optimal solution and let
λ1 (b∗ ), ..., λk (b∗ ) be the corresponding Lagrange multipliers. Suppose that as
b varies near b∗ , then x∗1 (b∗ ), ..., x∗n (b∗ ) and λ1 (b∗ ), ..., λk (b∗ ) are differentiable
functions and that x∗ (b∗ ) satisfies the constraint qualification. Then for each
j = 1, 2, ..., k :
∂
λj (b∗ ) = f (x∗ (b∗ ))
∂bj
• Proof: For simplicity, we do here the case of a single equality constraint, and
with f and g being functions of two variables. The Lagrangian is
36
The solution satisfies:
∂L ∗
0 = (x (b), y ∗ (b), λ∗ (b); b)
∂x
∂f ∗ ∂h
= (x (b), y ∗ (b)) − λ∗ (b) (x∗ (b), y ∗ (b), λ∗ (b)),
∂x ∂x
∂L ∗
0 = (x (b), y ∗ (b), λ∗ (b); b)
∂y
∂f ∗ ∂h
= (x (b), y ∗ (b)) − λ∗ (b) (x∗ (b), y ∗ (b), λ∗ (b)),
∂y ∂y
• Recall that
L(x, y, λ) = f (x, y) − λ(g(x, y) − b),
So that
d ∂
f (x(b), y(b); b) = λ(b) = L(x(b), y(b), λ(b); b)
db ∂b
37
Hence, what we have found above is simply a particular case of the envelope
theorem, which says that
d ∂
f (x(b), y(b); b) = L(x(b), y(b), λ(b); b)
db ∂b
Let x∗1 (c), ..., x∗n (c) denote the optimal solution and let µ1 (c), ..., µk (c) be the
corresponding Lagrange multipliers. Suppose that x∗1 (c), ..., x∗n (c) and µ1 (c),
..., µk (c) are differentiable functions and that x∗ (c) satisfies the constraint qual-
ification. Then for each j = 1, 2, ..., k :
d ∂
f (x∗ (c); c) = L(x∗ (c), µ(c); c)
dc ∂c
• Note: if hi (x1 , x2 , ..., xn , c) = 0 can be expressed as some h0i (x1 , x2 , ..., xn )−c = 0,
then we are back at the previous case, in which we have found that
d ∂
f (x∗ (c), c) = L(x∗ (c), µ(c); c) = λj (c)
dc ∂c
• We will prove this for the simple case of an unconstrained problem. Let φ(x; a)
be a continuous function of x ∈ Rn and the scalar a. For any a,consider the
maximization problem of max φ(x; a). Let x∗ (a) be the solution of this problem
and a continuous and differentiable function of a. We will show that
d ∂
φ(x∗ (a); a) = φ(x∗ (a); a)
da ∂a
38
We compute via the chain rule that
d X ∂φ ∂x∗ ∂φ ∗
φ(x∗ (a); a) = (x∗ (a); a) i (a) + (x (a); a)
da i
∂x i ∂a ∂a
∂φ ∗
= (x (a); a)
∂a
∂φ ∗
since (x (a); a) = 0 for all i by the first order conditions.
∂xi
• Intuitively, when we are already at a maximum, changing slightly the parame-
ters of the problem or the constraints, does not affect the value through changes
∂φ ∗
in the solution x∗ (a), because (x (a); a) = 0.
∂xi
• When we use the envelope theorem we have to make sure though that we do
not jump to another solution in a discrete manner.
39
Comparative Statics
We have been using the Implicit Function Theorem (IFT) throughout without
stating and explaining why we can use it. The IFT allows us to be assured that a set
of simultaneous equations:
y1 = f 1 (x1 , ..., xm )
y2 = f 2 (x1 , ..., xm )
..
.
yn = f n (x1 , ..., xm )
In other words, what the conditions of the IFT serve to do is to assure that the n
equations can in principle be solved for the n variables, y1 , ..., yn , even if we may not
be able to obtain the solution in an explicit form.
40
• Given the set of simultaneous equations above, if the functions F 1 , .., F n all
have continuous partial derivatives with respect to all x and y variables, and if
at a point (y0 , x0 ) that solves the set of simultaneous equations the determinant
of the (n × n) Jacobian w.r.t. the y-variables is not 0:
∂F 1 ∂F 1
... ∂F 1
∂y1 ∂y2 ∂yn
∂F 2 ∂F 2 ∂F 2
...
|J| = ∂y1 ∂y2 ∂yn
6= 0
... ... ...
∂F n ∂F n n
∂y1 ∂y2
... ∂F
∂yn
• It is then possible to find the partial derivatives of the implicit functions without
having to solve them for the y variables. Taking advantage of the fact that in
the neighborhood of the solution, the set of equations have a status of identities,
we can take the total differential of each equation and write dF j = 0. When
considering only dx1 6= 0 and setting the rest dxi = 0, the result, in matrix
notation, is (we will go through an example later in class):
∂F 1 ∂F 1 ∂F 1 ∂F 1
∂y1
∂y1 ∂y2
... ∂yn ∂x1 ∂x1
∂F 2 ∂F 2 ∂F 2 ∂y2 ∂F 2
∂y1 ∂y2
... ∂yn
∂x1
= − ∂x1
... ...
...
... ...
∂F n ∂F n ∂F n ∂yn ∂F n
∂y1 ∂y2
... ∂yn ∂x1 ∂x1
41
• Finally, since |J| is non zero there is a unique nontrivial solution to this linear
system, which by Cramer’s rule can be identified in the following way:
∂yj |Jj|
= .
∂x1 |J|
This is for general problems. Optimization problems have a unique feature: the
condition that indeed |J| =
6 0. (What is J? it is simply the matrix of partial
second derivatives of L, or what we call the bordered Hessian). We will see that
later on.
This means that indeed we can take the maximum value function, or set of
equilibrium conditions, totally differentiate them and find how the endogenous
variables change with the exogenous ones in the neighbourhood of the solution.
For example, for the case of optimization with one equality constraint:
F 1 (λ, x, y; b) = 0
F 2 (λ, x, y; b) = 0
F 3 (λ, x, y; b) = 0
is given by
b − g(x, y) = 0
fx − λgx = 0
fy − λgy = 0
We need to ensure that the Jacobian is not zero and then then we can use total
differentiation.
42
Coming back to the condition about the Jacobian, we need to ensure that:
∂F 1 ∂F 1 ∂F 1
∂λ ∂x ∂y
∂F 2 ∂F 2 ∂F 2
|J| = ∂λ ∂x ∂x
6= 0
∂F 3 ∂F 3 ∂F 3
∂λ ∂x ∂y
or:
0 −gx −gy
−gx fxx − λgxx fxy − λgxy 6= 0
−gy fxy − λgxy fyy − λgyy
but the determinant of J, is that of the bordered Hessian H̄. Whenever sufficient
second order conditions are satisfied, we know that the determinant of the bordered
Hessian is not zero (in fact it is positive).
Now we can totally differentiate the equations:
gx dx + gy dy − 1db = 0
(fxx − λgxx )dx + (fxy − λgxy )dy − gx dλ = 0
(fyx − λgyx )dx + (fyy − λgyy )dy − gy dλ = 0
∂x ∂y ∂λ
where at the equilibrium solution, one can then solve for , , .
∂b ∂b ∂b
43