
Heijdra Foundations of Modern Macroeconomics (Oxford, 2002)
.pdfefined as the transposed
(A.15)
It.
unique inverse, denoted
(A.16)
= I.
non-singular matrix)
pute AA-1 and A -1 A
Mathematical Appendix
Assuming that the indicated inverses exist (and the matrices A and B are thus non-singular), we find the following properties:
(A T)-1 = (A -1 ) T |
(A.17) |
|
(AB) -1 =B-1 A-1 |
||
114-1 1=1A1 -1 |
|
|
I-1 |
=1 |
|
(A-1 )-1 |
=A |
|
A.2.5 Cramer's Rule
Suppose we have a linear system of n equations in n unknowns:
anxi + a12x2 + " • + ainxn =
anxi + a22X2 + • • • + a2nxn = b2
(A.18)
anixi + an2x2 + • • • + annxn =
where au are the coefficients, bi are the exogenous variables, and xi are the endogenous variables. We can write this system in the form of a single matrix equation as:
Ax = b, |
|
|
|
|
|
|
|
|
(A.19) |
where A is an n by n matrix, and x and b are n by 1 (column) vectors: |
|
||||||||
all |
a12 |
|
• |
ain |
|
— |
|
bi |
|
|
|
X1 |
|
|
|||||
a21 |
a22 |
• |
• |
a2n |
|
X2 |
|
b2 |
|
A - |
|
|
|
|
, X |
|
, b |
|
(A.20) |
and |
an2 • |
• |
ann |
|
_ xn _ |
|
bn |
|
Provided the coefficient matrix A is non-singular (so that IA l 0 0) the solution of the matrix equation is:
x = |
(A.21) |
Instead of inverting the entire matrix A we can find the solutions for individual variables by means of Cramer's Rule (which only involves determinants):
x . _ |
(A.22) |
— IA I , for j = 1, 2, , n |
Mathematical Appendix
where 1A1 1 is the determinant of the matrix A1 which is obtained by replacing column j of A by the vector of exogenous variables, for example A 1 is:
b1 |
a12 |
a13 |
• • • |
ain |
b2 |
a22 |
a23 |
• |
a2n |
Al |
|
|
|
(A.23) |
_ bn |
ant |
an3 |
•• • |
ann _ |
If the vector b consists entirely of zeros we call the system homogeneous. If IA I 0 0 then the unique solution to the matrix equation is the trivial one: x = A -1 b = 0. The only way to get a non-trivial solution to a homogeneous system is if the coefficient matrix is singular, i.e. if IA I = 0. In that case Cramer's Rule cannot be used. An infinite number of solutions nevertheless exist (including the trivial one) in that case. Take, for example, the following homogeneous system:
1 2 xi |
0 |
(A.24) |
|
[Z 4 [ x2 = [ 0 |
|||
|
Clearly, AlI= 4 — 4 = 0 so the system is singular (row 2 is two times row 1). Nevertheless, both the trivial solution (x 1 = x2 = 0) and an infinite number of nontrivial solutions (any combination for which x 1 + 2x2 = 0) exist. Intuitively, we have infinitely many solutions because we have a single equation but two unknowns.
A.2.6 Characteristic roots and vectors
A characteristic vector of an n by n matrix A is a non-zero vector x which, when premultiplied by A yields a multiple of the same vector:
Ax = Ax, |
(A.25) |
where A is called the characteristic root (or eigenvalue) of A. By rewriting equation (A.25) we find:
(Al — A)x = 0, |
(A.26) |
which constitutes a homogeneous system of equations which has non-trivial solutions provided the determinant of its coefficient matrix, AI — A, is zero:
IA I —Al --= 0 |
(A.27) |
This expression is called the characteristic equation of A. For a 2 by 2 matrix the characteristic equation can be written as:
IA — =[ |
A — an |
a12 |
= — ail) (A. — a22) — ai2a2i |
|
a21 |
A — a22 |
|
||
|
|
|
||
= A2 — (an + a22)), + a11a22 — a12a21 |
|
|||
= A2 tr(A)A + AlI= 0, |
|
(A.28) |
where tr(A) and IA l are. I Hence, for such a matro possesses two roots:
tr(A) |
8/[tr(.:i |
A1,2 — |
2 1 |
|
These roots are distinct real (rather than comply IA < 0). For an n by n with n roots, A1 , A2, characteristic roots are:
Xi = tr(A) 117_1 Ai = IA
Associated with each chan is unique up to a const (A.26). If a matrix has follows:
P-1AP = A <#. A
where P is the matrix \\ diagonal matrix with chui tion is useful in the conte and below.
Intermezzo
Eigenvalues, eigenvt.‘ defined as:
A |
6 |
10 |
|
—2 —3 |
|||
|
The characteristic eq, characteristic roots ar. with Al is obtained b\
—410 [
664
"Pd by replacing column
(A.23)
homogeneous. If IA 1 0 0 I one: x = A -l b = 0. The c'em is if the coefficient cannot be used. An the trivial one) in that
1.
(A.24)
2 is two times row 1). finite number of noncist. Intuitively, we have n but two unknowns.
vector x which, when
I
(A.25)
By rewriting equation
(A.26)
-hich has non-trivial c, - A, is zero:
(A.27)
r a 2 by 2 matrix the
• -a21
(A.28)
Mathematical Appendix
where tr(A) and IA 1 are, respectively, the trace and the determinant of matrix A. Hence, for such a matrix the characteristic equation is quadratic in A and thus possesses two roots:
A1,2 = |
tr(A) +.1[tr(A)] 2 - 4 IA1 |
(A.29) |
|
2 |
|||
|
|
These roots are distinct if the discriminant, [tr(A)] 2 - 41A1, is non-zero. They are real (rather than complex) if the discriminant is positive (this is certainly the case if 1A 1 < 0). For an n by n matrix the characteristic equation is an n-th order polynomial with n roots, —1, —2, • • • , An which may not all be distinct or real. Some properties of characteristic roots are: -
Eri' 1 Ai = tr(A) |
(A.30) |
|
r17-1 Ai = 1A1 |
||
|
Associated with each characteristic root is a chara eristic ve or (or eigenvector), which is unique up to a constant. The characteristic vector x(i) associated with Xi solves
(A.26). If a matrix has distinct characteristic roots then it can be |
diagonalized as |
follows: |
|
P-1AP = A <4. A = PAP-1 , |
(A.31) |
where P is the matrix with the characteristic vectors, x(i) , as columns and A is the diagonal matrix with characteristic roots, X i , on the principal diagonal. Diagonalization is useful in the context of difference and differential equations—see Chapter 2 and below.
Intermezzo
Eigenvalues, eigenvectors, and matrix diagonalization. Suppose that A is defined as:
A = 6 10 -2 -3
The characteristic equation is 2, 2 — 3A + 2 = - 1)(A -- 2) = 0 so that the characteristic roots are A-= 1 and 2,2 = 2. The characteristic vector associated with Al is obtained by noting from (A.26) that:
(A 11-A)x = 0
([ 01 01 [ 26103 1) [
665
Mathematical Appendix
Any solution for which 2x1 + 4x2 0 will do. Hence, by setting x1 = c (a non-zero constant) we find that x2 = -c/2 so that the characteristic vector associated with Xi is:
x( )[ C -c/2 •
Similarly, for A2 = 2 we find:
(A.21 A)x -= 0 |
|
|
|
|
2 |
0 |
6 |
10 |
x1 |
0 |
2 |
-2 |
-3 1) |
X2 |
-4 -10 |
[ Xi |
0 |
1• |
|
2 |
5 |
X2 |
0 |
Any combination for which 2x1 + 5x2 = 0 will do. Hence, the characteristic vector associated with
X(2)
-2c/5
In the example matrix we have:
-c/2 -2c/5 |
and |
A |
1 |
0 |
|
0 |
2 |
||||
|
|
from which we verify the result:
PAP-1 .--- 10 [ |
cc |
1 |
0 I |
|
-2c/5 |
-c |
|
|
|
c2 -c / 2 -2c/5 |
0 |
2 |
|
c/2 |
c |
|
|
||
=10 |
1 |
1 |
1 0 1 |
-2/5 - |
|
|
|
||
-1/2 |
-2/5 |
0 2_ 1/2 |
|
|
|
||||
|
|
|
|
||||||
= 10 |
1 |
2 |
-2/5 - 1 |
= 10 |
3/5 |
1 |
=- A |
||
|
-1/2 |
-4/5 |
1/2 |
|
1 |
-1/5 -3/10 |
|
It works!
A.2.7 Literature
Basic: Klein (1998, chs. 4-5), Chiang (1984, chs. 4-5), Sydsxter and Hammond (1995, chs. 12-14). Intermediate: Intriligator (1971, appendix B), Kreyszig (1999,
chs. 6-7), and Strang (19 (1985), and Ortega (19b
I
A.3 Implicit Funct
A.3.1 Single equation
Suppose we have the fi interest, y, to one or mor
F(y, xi, .. xm) =
Assume that (a) F has co aF/axi for j = 1,2, ... ,n (A.32). Then according - neighbourhood of [x ) ] variables:
Y = f (Xl, X2, • • • Xrr:'
I
The implicit function is c by fj af /axe, which c.
ay = f = - L,t . ax, F
As an example, consider that ay/ax = -Fx /Fy = -
A.3.2 System of equa
Next we consider the sysi
I
Fl (yi, y2, yn; xi,.
F2 (yi, y2, • • • , yn;
Fn (yi, y2, • • Yn; xl
I
We assume that (a) th. respect to all yi and xi ai
666
v setting xi c (a characteristic vector
Mathematical Appendix
chs. 6-7), and Strang (1988). Advanced: Ayres (1974), Lancaster and Tismenetsky (1985), and Ortega (1987).
A.3 Implicit Function Theorem
A.3.1 Single equation
Suppose we have the following equation relating the endogenous variable of interest, y, to one or more/exogenous variables, xi:
|
|
|
|
F(y, x1, x2, • • • , Xm) = 0. |
|
(A.32) |
||
|
|
|
Assume that (a) F has continuous partial derivatives (denoted by Fy aF /ay, Fi |
|||||
|
|
|
aF /axi for j = 1, 2, , m) and (b) Fy 0 0 around a point [y°, x?] which satisfies |
|||||
ce, the characteristic |
(A.32). Then according to the implicit function theorem, there exists an m-dimensional |
|||||||
|
|
|
neighbourhood of [x1]( in which y is an implicitly defined function of the exogenous |
|||||
|
|
|
variables: |
|
|
|
||
|
|
|
|
|
|
|
(A.33) |
|
|
|
|
The implicit function is continuous and has continuous partial derivatives, denoted |
|||||
|
|
|
by fi |
f/axi, which can be computed as follows: |
||||
|
|
' |
|
ay c |
for i — |
1 I • • • I tn• |
(A.34) |
|
|
|
|
a= " = F |
|||||
|
|
|
|
' |
|
|
||
-c |
|
|
As an example, consider F(y, x) = y2 + x2 - 9. We find that Fy = 2y and FX = 2x, so |
|||||
|
|
|
that ay/ax = -Fx /Fy = -x/y provided y 0. |
|
||||
1 |
|
|
|
|
|
|
|
|
3/5 |
1 |
|
A.3.2 System of equations |
|
|
|||
= A |
Next we consider the system of n equations in n endogenous variables (Y1, y2, • • • , Yn): |
|||||||
-1/5 |
-3/10 |
|||||||
|
|
|
|
|
|
|||
|
|
|
|
F 1 (Y1, Y2, • • • , |
xl, x2, • . Xm ) = 0 |
|
||
|
|
|
|
F2 (yi, y2, • • • |
Yn; xl, X2, • . , X m ) = 0 |
(A.35) |
||
|
|
|
|
|
|
|
||
|
|
|
|
Fn (yi, Y2,•• • I Yn; Xi, X2, • . Xm ) = 0 |
|
|||
iydsxter and Hammond |
We assume that (a) the functions F' all have continuous partial derivatives with |
|||||||
,lix B), Kreyszig (1999, |
respect to all yi and xi and (b) at a point [4; x1]( |
the following determinant (of the |
||||||
|
|
|
|
|
|
|
667 |
Mathematical Appendix
Jacobian matrix) is non-zero: |
|
|
|
aFl /ay2 |
aFl/ayn |
|
|
aF2 /ayi aF2 /0y2 |
aF2 /ayn |
O. |
(A.36) |
I/1 |
|
||
aFn/ayi Un/ay2 • • |
• aFn/ayn |
|
|
Then, according to the generalized implicit function theorem there exists an m- dimensional neighbourhood of [4] in which the variables yi are implicitly defined functions of the exogenous variables:
yi= fl (X1, x2, . |
x.) |
|
y2 = f2 (xi , x2, |
x.) |
(A.37) |
: = |
|
|
|
|
|
yn = fn(xi, x2, |
x.) |
|
These implicit functions are continuous and have continuous partial derivatives, denoted by afiiaxi, which can be computed as follows:
aY! |
|
1! |
for i = 1, 2, • • • , n, |
(A.38) |
—I |
||||
ax; |
= fi |
I/1 |
|
|
where J! is the matrix obtained by replacing column i of matrix J by the following vector of partial derivatives:
--8F1 /axi -
-0F2 /ax;
(A.39)
-81'n/ax1 -
Intermezzo
Generalized implicit function theorem. As a example, consider the IS-LM model:
Y = C(Y - T(Y)) + .1(r) + Go
Mo L(r, Y),
where Y is output, C is consumption, I is investment, r is the interest rate, T is taxes. The endogenous variables are r and Y and the exogenous variables are government consumption Go and the money supply Mo. By differentiating
with respect to Go v L
1 - Cy_T(1 -
Ly
The Jacobian determin
Ill Lr [1 - CY-7
where the sign folios% depend negatively on ti sity to consume and th
0 < Cy_T(1 -- Ty) <
(Ly > 0). By Cramer's F
aY |
1 |
1 -I, |
aGo |
|
0 I |
ar |
1 |
1 - |
aGo |
I/1 |
|
These expressions, o: _
A.3.3 Literature
Basic: Klein (1998, pp. 2 a (1995, pp. 591-593). Adv
A.4 Static Optimiz
A.4.1 Unconstrainedi
Suppose we wish to fn. function:
Y = f (X),
where we assume that th tives. The necessary con
668
(A.36)
qz there exists an m- v are implicitly defined
(A.37)
ous partial derivatives,
(A.38)
• i x I by the following
(A.39)
sider the IS-LM
the interest rate, xogenous variables differentiating
Mathematical Appendix
with respect to Go we get:
Cy-T(1 - Ty)
Ly Lr
The Jacobian determinant is:
Lr [1 - CY-T(1 - TY)} + IrL < 0,
where the sign follows from the fact that both money demand and investment depend negatively on the interest rate 0 and Ir < 0), the marginal propensity to consume and the marginal tax rate are between zero and unity (so that 0 < Cy_T(1 -- Ty) < 1), and money demand depends positively on output
0). By Cramer's Rule we get the partial derivatives:
a Y |
1 |
8G0 |
|
ar |
1 |
aG0 |
Ill |
These expressions, of course, accord with intuition (see Chapter 1).
A.3.3 Literature
Basic: Klein (1998, pp. 239-245), Chiang (1984, ch. 8), and Sydsxter and Hammond (1995, pp. 591-593). Advanced: De la Fuente (2000, ch. 5).
A.4 Static Optimization
A.4.1 Unconstrained optimization
Suppose we wish to find an optimum (minimum or maximum) of the following function:
= f (A), |
(A.40) |
where we assume that this function is continuous and possesses continuous derivatives. The necessary condition for a (relative) extremum of the function at point
669
Mathematical Appendix
X = X0 is
f (xo) = 0. |
(A.41) |
To test whether f(x) attains a relative maximum or a relative minimum at x = xo we compute the second derivative. The second-order sufficient condition is:
if f"(xo) |
0, f (x0) is a relative { maximum |
(A.42) |
|
minimum |
|
Now suppose that the function depends on n arguments (choice variables): |
||
y = f(xl, x2, . , xn), |
(A.43) |
where f(.) is continuous and possesses continuous derivatives. The first-order necessary conditions for a relative extremum are:
— 0, i = 1, 2, • • • , n, |
(A.44) |
|
where f |
f/axi are the partial derivatives of f 0 with respect to Xi. To study |
the second-order sufficient conditions we define the Hessian matrix of second-order derivatives, H:
|
fll f12 |
|
• |
fin |
|
Hn.n |
f21 f22 |
• |
• • • |
f2n |
(A.45) |
|
|
|
|
||
|
_ fni fn2 • • • • • • fnn _ |
|
|||
where fii |
0 2084' and fii |
|
a2 f/axiaxi are second-order partial derivatives. By |
Young's theorem we know that fii = fii so the Hessian matrix is symmetric. We define the following set of principal minors of H:
|
|
|
|
fll fi2 |
fin |
|
fu f12 |
|
111,1 |
f21 f22 • |
• • • f2n |
11111 |
11, H2I= f21 f22 |
• •, |
|
|
|
|
|
|
|||
|
|
|
|
fnl fn2 • • • |
fnn |
|
|
|
|
|
(A.46) |
Then, provided the first-order conditions hold at a point [x7, x°, .., x°], the secondorder sufficient condition for f (4) to be a relative maximum is:
lib I < 0, I H2 I > 0, . (— IHnl > 0, (A.47)
whilst for a relative minimum the condition is:
11121, |
111, 1 > 0 |
(A.48) |
See Chiang (1984, pp. 337-353) for the relation between concavity—convexity of f 0 and the second-order conditions.
A.4.2 Equality constra
We focus on the case ‘.. straint. As in the unconst constraint is given by:
g(Xlt X2, • • • , Xn) |
C, |
where c is a constant. \\ L derivatives. The Lagran,
f(xl,x2, |
, xn) 4 |
where A is the Lagran0 _ extremum are:
Li = 0, i = 1,
LA = 0,
670
(A.41)
minimum at x xo ,-gt condition is:
(A.42)
(choice variables):
I |
(A.43) |
|
tives. The first-order
(A.44)
respect to xi. To study t matrix of second-order
(A.45)
partial derivatives. By -frix is symmetric. We
' • fin
" f2n
|
|
fnn |
|
|
(A.46) |
A o |
o |
], the second- |
. x,, |
xn |
is:
(A.47)
(A.48)
7nncavity—convexity of
Mathematical Appendix
A.4.2 Equality constraints
We focus on the case with multiple choice variables and a single equality constraint. As in the unconstrained case, the objective function is given by (A.43). The constraint is given by:
g(xi, x2, |
, xn ) = c, |
(A.49) |
where c is a constant. We assume that g(.) is continuous and possesses continuous derivatives. The Lagrangian is defined as follows:
L f(xl,x2, . , xn ) + [c — g(xi, x2, . , xn)] |
(A.50) |
where A is the Lagrange multiplier. The first-order necessary conditions for an extremum are:
(A.51)
LA = 0,
where A aL/axi and LA aL/ax are the partial derivatives of the Lagrangean with respect to xi and A, respectively. To study the second-order conditions we formulate a so-called bordered Hessian matrix, denoted by H:
o |
gi g2 |
•• • gn |
|
gl |
fi2 • |
• fin |
(A.52) |
An+1)x (n+1) ==."' g2 |
f21 f22 • • |
• " f2n |
_ gn fnl fn2 " • • • • fnn _
The bordered Hessian consists of the ordinary Hessian but with the borders made up of the derivatives of the constraint function (gi). We define the following set of principal minors of H:
|
|
|
0 gl g2 |
• • • |
gn |
|
|
gl g2 |
|
f12 " |
• ' • |
fin |
|
|
IHn I |
g2 f2i f22 • • |
|
f2n . |
(A.53) |
|
H2 |
gl fn f12 |
• • |
||||
|
g2 f2i f22 |
|
|
|
|
|
|
|
|
gn fn 1 fn2 • |
• • • fnn |
|
|
Then provided the first-order conditions hold at a point [x?, |
|
, x°„] the second- |
||||
order sufficient conditions for f (x?) to be a relative constrained maximum are: |
||||||
( — 1)k Ifik I > 0, k = 2, ... , n, |
|
|
|
|
(A.54) |
whilst for the conditions for a relative constrained minimum are:
Ifik < |
k = 2, ... , n. |
(A.55) |
If there are multiple constraints then additional Lagrange multipliers are added to the Lagrangian (one per constraint) and the first-order condition for each Lagrange
671
Mathematical Appendix
multiplier, Al , takes the form LAi aLiaa.; = 0. See Chiang (1984, pp. 385-386) for the appropriately defined bordered Hessian for the multi-constraint case.
Interpretation of the Lagrange multiplier
We now return to the single constraint case in order to demonstrate the interpretation of the Lagrange multiplier in the optimum. Using the superscript "0" to denote optimized values, we can write the optimized value of the Lagrangian as:
LO -C I 0 0 |
, |
0‘ , 0 |
, ..,x)]. |
(A.56) |
X2 |
, Xn ) + A [C. — g(x7, x2() |
Next, we ask the question what happens if the constraint is changed marginally. Obviously, both A,13 and x? are expected to change if c does. Differentiating (A.56) we get:
dr° |
n |
dx9 |
d)° |
xo (dc) = )1/4,0, |
(A.57) |
|
dc |
Li |
1 |
) + |
dc |
dc |
|
i=1 |
dc |
|
|
|||
|
|
|
|
|
|
where we have used the necessary conditions for an optimum (CA = Li = 0 for i = 1, 2, , n) to get from the first to the second equality. Recall that the constraint holds with equality (c = g(.)) so that A° measures the effect of a small change in c on the optimized value of the objective function f (.). For example, if the objective function is utility and c is income, then A° is the marginal utility of income.
A.4.3 Inequality constraints
We now briefly study some key results from non-linear programming. We first look at the simplest case with non-negativity constraints on the choice variables. Then we take up the more challenging case of general inequalities. We focus on first-order conditions and ignore some of the subtleties involved (like constraint qualifications and second-order conditions).
Non- negativity constraints
Suppose that the issue is to maximize a function y = f (x) subject only to the nonnegativity constraint x > 0. There are three situations which can arise. These have been illustrated in Figure A.1 which is taken from Chiang (1984, p. 723).
Panel (a) shows the case we have studied in detail above. The function attains a maximum for a strictly positive value of x. We call this an interior solution because the solution lies entirely within the feasible region (and not on a boundary). The constraint x > 0 is non-binding and the first-order condition is as before:
f'(xo) = 0. (interior solution)
Panels (b) and (c) deal with two types of boundary solutions. In panel (b) the function happens to attain a maximum for x = xo = 0, i.e. exactly on the boundary of the
feasible region. In panel
f'(xo) =-- 0 and I
Finally, in panel (c) we al f (x) continues to rise have:
f' (xo) 0 and
I These three conditions,
solutions, can be combi I
f' (x0) 0, xo
There are two key things A.1, we can safely ex, _
672