Добавил:
Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:

Heijdra Foundations of Modern Macroeconomics (Oxford, 2002)

.pdf
Скачиваний:
761
Добавлен:
22.08.2013
Размер:
27.75 Mб
Скачать

efined as the transposed

(A.15)

It.

unique inverse, denoted

(A.16)

= I.

non-singular matrix)

pute AA-1 and A -1 A

Mathematical Appendix

Assuming that the indicated inverses exist (and the matrices A and B are thus non-singular), we find the following properties:

(A T)-1 = (A -1 ) T

(A.17)

(AB) -1 =B-1 A-1

114-1 1=1A1 -1

 

I-1

=1

 

(A-1 )-1

=A

 

A.2.5 Cramer's Rule

Suppose we have a linear system of n equations in n unknowns:

anxi + a12x2 + " • + ainxn =

anxi + a22X2 + • • • + a2nxn = b2

(A.18)

anixi + an2x2 + • • • + annxn =

where au are the coefficients, bi are the exogenous variables, and xi are the endogenous variables. We can write this system in the form of a single matrix equation as:

Ax = b,

 

 

 

 

 

 

 

 

(A.19)

where A is an n by n matrix, and x and b are n by 1 (column) vectors:

 

all

a12

 

ain

 

 

bi

 

 

 

X1

 

 

a21

a22

a2n

 

X2

 

b2

 

A -

 

 

 

 

, X

 

, b

 

(A.20)

and

an2

ann

 

_ xn _

 

bn

 

Provided the coefficient matrix A is non-singular (so that IA l 0 0) the solution of the matrix equation is:

x =

(A.21)

Instead of inverting the entire matrix A we can find the solutions for individual variables by means of Cramer's Rule (which only involves determinants):

x . _

(A.22)

IA I , for j = 1, 2, , n

Mathematical Appendix

where 1A1 1 is the determinant of the matrix A1 which is obtained by replacing column j of A by the vector of exogenous variables, for example A 1 is:

b1

a12

a13

• • •

ain

b2

a22

a23

a2n

Al

 

 

 

(A.23)

_ bn

ant

an3

•• •

ann _

If the vector b consists entirely of zeros we call the system homogeneous. If IA I 0 0 then the unique solution to the matrix equation is the trivial one: x = A -1 b = 0. The only way to get a non-trivial solution to a homogeneous system is if the coefficient matrix is singular, i.e. if IA I = 0. In that case Cramer's Rule cannot be used. An infinite number of solutions nevertheless exist (including the trivial one) in that case. Take, for example, the following homogeneous system:

1 2 xi

0

(A.24)

[Z 4 [ x2 = [ 0

 

Clearly, AlI= 4 — 4 = 0 so the system is singular (row 2 is two times row 1). Nevertheless, both the trivial solution (x 1 = x2 = 0) and an infinite number of nontrivial solutions (any combination for which x 1 + 2x2 = 0) exist. Intuitively, we have infinitely many solutions because we have a single equation but two unknowns.

A.2.6 Characteristic roots and vectors

A characteristic vector of an n by n matrix A is a non-zero vector x which, when premultiplied by A yields a multiple of the same vector:

Ax = Ax,

(A.25)

where A is called the characteristic root (or eigenvalue) of A. By rewriting equation (A.25) we find:

(Al — A)x = 0,

(A.26)

which constitutes a homogeneous system of equations which has non-trivial solutions provided the determinant of its coefficient matrix, AI — A, is zero:

IA I —Al --= 0

(A.27)

This expression is called the characteristic equation of A. For a 2 by 2 matrix the characteristic equation can be written as:

IA — =[

A — an

a12

= — ail) (A. — a22) — ai2a2i

 

a21

A — a22

 

 

 

 

= A2 (an + a22)), + a11a22 — a12a21

 

= A2 tr(A)A + AlI= 0,

 

(A.28)

where tr(A) and IA l are. I Hence, for such a matro possesses two roots:

tr(A)

8/[tr(.:i

A1,2 —

2 1

 

These roots are distinct real (rather than comply IA < 0). For an n by n with n roots, A1 , A2, characteristic roots are:

Xi = tr(A) 117_1 Ai = IA

Associated with each chan is unique up to a const (A.26). If a matrix has follows:

P-1AP = A <#. A

where P is the matrix \\ diagonal matrix with chui tion is useful in the conte and below.

Intermezzo

Eigenvalues, eigenvt.‘ defined as:

A

6

10

—2 —3

 

The characteristic eq, characteristic roots ar. with Al is obtained b\

410 [

664

"Pd by replacing column

(A.23)

homogeneous. If IA 1 0 0 I one: x = A -l b = 0. The c'em is if the coefficient cannot be used. An the trivial one) in that

1.

(A.24)

2 is two times row 1). finite number of noncist. Intuitively, we have n but two unknowns.

vector x which, when

I

(A.25)

By rewriting equation

(A.26)

-hich has non-trivial c, - A, is zero:

(A.27)

r a 2 by 2 matrix the

-a21

(A.28)

Mathematical Appendix

where tr(A) and IA 1 are, respectively, the trace and the determinant of matrix A. Hence, for such a matrix the characteristic equation is quadratic in A and thus possesses two roots:

A1,2 =

tr(A) +.1[tr(A)] 2 - 4 IA1

(A.29)

2

 

 

These roots are distinct if the discriminant, [tr(A)] 2 - 41A1, is non-zero. They are real (rather than complex) if the discriminant is positive (this is certainly the case if 1A 1 < 0). For an n by n matrix the characteristic equation is an n-th order polynomial with n roots, —1, —2, • • • , An which may not all be distinct or real. Some properties of characteristic roots are: -

Eri' 1 Ai = tr(A)

(A.30)

r17-1 Ai = 1A1

 

Associated with each characteristic root is a chara eristic ve or (or eigenvector), which is unique up to a constant. The characteristic vector x(i) associated with Xi solves

(A.26). If a matrix has distinct characteristic roots then it can be

diagonalized as

follows:

 

P-1AP = A <4. A = PAP-1 ,

(A.31)

where P is the matrix with the characteristic vectors, x(i) , as columns and A is the diagonal matrix with characteristic roots, X i , on the principal diagonal. Diagonalization is useful in the context of difference and differential equations—see Chapter 2 and below.

Intermezzo

Eigenvalues, eigenvectors, and matrix diagonalization. Suppose that A is defined as:

A = 6 10 -2 -3

The characteristic equation is 2, 2 3A + 2 = - 1)(A -- 2) = 0 so that the characteristic roots are A-= 1 and 2,2 = 2. The characteristic vector associated with Al is obtained by noting from (A.26) that:

(A 11-A)x = 0

([ 01 01 [ 26103 1) [

665

A2 is:

Mathematical Appendix

Any solution for which 2x1 + 4x2 0 will do. Hence, by setting x1 = c (a non-zero constant) we find that x2 = -c/2 so that the characteristic vector associated with Xi is:

x( )[ C -c/2 •

Similarly, for A2 = 2 we find:

(A.21 A)x -= 0

 

 

 

2

0

6

10

x1

0

2

-2

-3 1)

X2

-4 -10

[ Xi

0

1

2

5

X2

0

Any combination for which 2x1 + 5x2 = 0 will do. Hence, the characteristic vector associated with

X(2)

-2c/5

In the example matrix we have:

-c/2 -2c/5

and

A

1

0

0

2

 

 

from which we verify the result:

PAP-1 .--- 10 [

cc

1

0 I

 

-2c/5

-c

 

 

c2 -c / 2 -2c/5

0

2

 

c/2

c

 

 

=10

1

1

1 0 1

-2/5 -

 

 

 

-1/2

-2/5

0 2_ 1/2

 

 

 

 

 

 

 

= 10

1

2

-2/5 - 1

= 10

3/5

1

=- A

 

-1/2

-4/5

1/2

 

1

-1/5 -3/10

 

It works!

A.2.7 Literature

Basic: Klein (1998, chs. 4-5), Chiang (1984, chs. 4-5), Sydsxter and Hammond (1995, chs. 12-14). Intermediate: Intriligator (1971, appendix B), Kreyszig (1999,

chs. 6-7), and Strang (19 (1985), and Ortega (19b

I

A.3 Implicit Funct

A.3.1 Single equation

Suppose we have the fi interest, y, to one or mor

F(y, xi, .. xm) =

Assume that (a) F has co aF/axi for j = 1,2, ... ,n (A.32). Then according - neighbourhood of [x ) ] variables:

Y = f (Xl, X2, • • • Xrr:'

I

The implicit function is c by fj af /axe, which c.

ay = f = - L,t . ax, F

As an example, consider that ay/ax = -Fx /Fy = -

A.3.2 System of equa

Next we consider the sysi

I

Fl (yi, y2, yn; xi,.

F2 (yi, y2, • • • , yn;

Fn (yi, y2, • • Yn; xl

I

We assume that (a) th. respect to all yi and xi ai

666

v setting xi c (a characteristic vector

Mathematical Appendix

chs. 6-7), and Strang (1988). Advanced: Ayres (1974), Lancaster and Tismenetsky (1985), and Ortega (1987).

A.3 Implicit Function Theorem

A.3.1 Single equation

Suppose we have the following equation relating the endogenous variable of interest, y, to one or more/exogenous variables, xi:

 

 

 

 

F(y, x1, x2, • • • , Xm) = 0.

 

(A.32)

 

 

 

Assume that (a) F has continuous partial derivatives (denoted by Fy aF /ay, Fi

 

 

 

aF /axi for j = 1, 2, , m) and (b) Fy 0 0 around a point [y°, x?] which satisfies

ce, the characteristic

(A.32). Then according to the implicit function theorem, there exists an m-dimensional

 

 

 

neighbourhood of [x1]( in which y is an implicitly defined function of the exogenous

 

 

 

variables:

 

 

 

 

 

 

 

 

 

 

(A.33)

 

 

 

The implicit function is continuous and has continuous partial derivatives, denoted

 

 

 

by fi

f/axi, which can be computed as follows:

 

 

'

 

ay c

for i

1 I • • • I tn

(A.34)

 

 

 

a= " = F

 

 

 

 

'

 

 

-c

 

 

As an example, consider F(y, x) = y2 + x2 - 9. We find that Fy = 2y and FX = 2x, so

 

 

 

that ay/ax = -Fx /Fy = -x/y provided y 0.

 

1

 

 

 

 

 

 

 

3/5

1

 

A.3.2 System of equations

 

 

= A

Next we consider the system of n equations in n endogenous variables (Y1, y2, • • • , Yn):

-1/5

-3/10

 

 

 

 

 

 

 

 

 

 

F 1 (Y1, Y2, • • • ,

xl, x2, • . Xm ) = 0

 

 

 

 

 

F2 (yi, y2, • • •

Yn; xl, X2, • . , X m ) = 0

(A.35)

 

 

 

 

 

 

 

 

 

 

 

Fn (yi, Y2,•• • I Yn; Xi, X2, • . Xm ) = 0

 

iydsxter and Hammond

We assume that (a) the functions F' all have continuous partial derivatives with

,lix B), Kreyszig (1999,

respect to all yi and xi and (b) at a point [4; x1](

the following determinant (of the

 

 

 

 

 

 

 

667

Mathematical Appendix

Jacobian matrix) is non-zero:

 

 

 

aFl /ay2

aFl/ayn

 

 

aF2 /ayi aF2 /0y2

aF2 /ayn

O.

(A.36)

I/1

 

aFn/ayi Un/ay2 • •

aFn/ayn

 

 

Then, according to the generalized implicit function theorem there exists an m- dimensional neighbourhood of [4] in which the variables yi are implicitly defined functions of the exogenous variables:

yi= fl (X1, x2, .

x.)

 

y2 = f2 (xi , x2,

x.)

(A.37)

: =

 

 

 

yn = fn(xi, x2,

x.)

 

These implicit functions are continuous and have continuous partial derivatives, denoted by afiiaxi, which can be computed as follows:

aY!

 

1!

for i = 1, 2, • • • , n,

(A.38)

I

ax;

= fi

I/1

 

 

where J! is the matrix obtained by replacing column i of matrix J by the following vector of partial derivatives:

--8F1 /axi -

-0F2 /ax;

(A.39)

-81'n/ax1 -

Intermezzo

Generalized implicit function theorem. As a example, consider the IS-LM model:

Y = C(Y - T(Y)) + .1(r) + Go

Mo L(r, Y),

where Y is output, C is consumption, I is investment, r is the interest rate, T is taxes. The endogenous variables are r and Y and the exogenous variables are government consumption Go and the money supply Mo. By differentiating

with respect to Go v L

1 - Cy_T(1 -

Ly

The Jacobian determin

Ill Lr [1 - CY-7

where the sign folios% depend negatively on ti sity to consume and th

0 < Cy_T(1 -- Ty) <

(Ly > 0). By Cramer's F

aY

1

1 -I,

aGo

 

0 I

ar

1

1 -

aGo

I/1

 

These expressions, o: _

A.3.3 Literature

Basic: Klein (1998, pp. 2 a (1995, pp. 591-593). Adv

A.4 Static Optimiz

A.4.1 Unconstrainedi

Suppose we wish to fn. function:

Y = f (X),

where we assume that th tives. The necessary con

668

(Ly >
(Lr <

(A.36)

qz there exists an m- v are implicitly defined

(A.37)

ous partial derivatives,

(A.38)

i x I by the following

(A.39)

sider the IS-LM

the interest rate, xogenous variables differentiating

Mathematical Appendix

with respect to Go we get:

Cy-T(1 - Ty)

Ly Lr

The Jacobian determinant is:

Lr [1 - CY-T(1 - TY)} + IrL < 0,

where the sign follows from the fact that both money demand and investment depend negatively on the interest rate 0 and Ir < 0), the marginal propensity to consume and the marginal tax rate are between zero and unity (so that 0 < Cy_T(1 -- Ty) < 1), and money demand depends positively on output

0). By Cramer's Rule we get the partial derivatives:

a Y

1

8G0

 

ar

1

aG0

Ill

These expressions, of course, accord with intuition (see Chapter 1).

A.3.3 Literature

Basic: Klein (1998, pp. 239-245), Chiang (1984, ch. 8), and Sydsxter and Hammond (1995, pp. 591-593). Advanced: De la Fuente (2000, ch. 5).

A.4 Static Optimization

A.4.1 Unconstrained optimization

Suppose we wish to find an optimum (minimum or maximum) of the following function:

= f (A),

(A.40)

where we assume that this function is continuous and possesses continuous derivatives. The necessary condition for a (relative) extremum of the function at point

669

If there are multiple co the Lagrangian (one pc
1Hkl < 0, k =
2, .
I
_ gn
The bordered Hessian co up of the derivatives o: principal minors of H:
0 gl S2 f
g2 f2i
Then provided the fi rst-t order sufficient conditini
( — 1)k 111k1 > 0, k= whilst for the condition!
H(n+1)x(n+1)
o gi g2
bordered Hes\
Li -= aLlaxi
where and respect to xi and A, respec a so-called
2, .
11

Mathematical Appendix

X = X0 is

f (xo) = 0.

(A.41)

To test whether f(x) attains a relative maximum or a relative minimum at x = xo we compute the second derivative. The second-order sufficient condition is:

if f"(xo)

0, f (x0) is a relative { maximum

(A.42)

 

minimum

 

Now suppose that the function depends on n arguments (choice variables):

y = f(xl, x2, . , xn),

(A.43)

where f(.) is continuous and possesses continuous derivatives. The first-order necessary conditions for a relative extremum are:

— 0, i = 1, 2, • • • , n,

(A.44)

where f

f/axi are the partial derivatives of f 0 with respect to Xi. To study

the second-order sufficient conditions we define the Hessian matrix of second-order derivatives, H:

 

fll f12

 

fin

 

Hn.n

f21 f22

• • •

f2n

(A.45)

 

 

 

 

 

_ fni fn2 • • • • • • fnn _

 

where fii

0 2084' and fii

 

a2 f/axiaxi are second-order partial derivatives. By

Young's theorem we know that fii = fii so the Hessian matrix is symmetric. We define the following set of principal minors of H:

 

 

 

 

fll fi2

fin

 

fu f12

 

111,1

f21 f22 •

• • • f2n

11111

11, H2I= f21 f22

• •,

 

 

 

 

 

 

 

 

 

fnl fn2 • • •

fnn

 

 

 

 

 

(A.46)

Then, provided the first-order conditions hold at a point [x7, x°, .., x°], the secondorder sufficient condition for f (4) to be a relative maximum is:

lib I < 0, I H2 I > 0, . (— IHnl > 0, (A.47)

whilst for a relative minimum the condition is:

11121,

111, 1 > 0

(A.48)

See Chiang (1984, pp. 337-353) for the relation between concavity—convexity of f 0 and the second-order conditions.

A.4.2 Equality constra

We focus on the case ‘.. straint. As in the unconst constraint is given by:

g(Xlt X2, • • • , Xn)

C,

where c is a constant. \\ L derivatives. The Lagran,

f(xl,x2,

, xn) 4

where A is the Lagran0 _ extremum are:

Li = 0, i = 1,

LA = 0,

670

(A.41)

minimum at x xo ,-gt condition is:

(A.42)

(choice variables):

I

(A.43)

 

tives. The first-order

(A.44)

respect to xi. To study t matrix of second-order

(A.45)

partial derivatives. By -frix is symmetric. We

' • fin

" f2n

 

 

fnn

 

 

(A.46)

A o

o

], the second-

. x,,

xn

is:

(A.47)

(A.48)

7nncavity—convexity of

Mathematical Appendix

A.4.2 Equality constraints

We focus on the case with multiple choice variables and a single equality constraint. As in the unconstrained case, the objective function is given by (A.43). The constraint is given by:

g(xi, x2,

, xn ) = c,

(A.49)

where c is a constant. We assume that g(.) is continuous and possesses continuous derivatives. The Lagrangian is defined as follows:

L f(xl,x2, . , xn ) + [c — g(xi, x2, . , xn)]

(A.50)

where A is the Lagrange multiplier. The first-order necessary conditions for an extremum are:

(A.51)

LA = 0,

where A aL/axi and LA aL/ax are the partial derivatives of the Lagrangean with respect to xi and A, respectively. To study the second-order conditions we formulate a so-called bordered Hessian matrix, denoted by H:

o

gi g2

•• • gn

 

gl

fi2

fin

(A.52)

An+1)x (n+1) ==."' g2

f21 f22 • •

• " f2n

_ gn fnl fn2 " • • • • fnn _

The bordered Hessian consists of the ordinary Hessian but with the borders made up of the derivatives of the constraint function (gi). We define the following set of principal minors of H:

 

 

 

0 gl g2

• • •

gn

 

 

gl g2

 

f12 "

• ' •

fin

 

 

IHn I

g2 f2i f22 • •

 

f2n .

(A.53)

H2

gl fn f12

• •

 

g2 f2i f22

 

 

 

 

 

 

 

 

gn fn 1 fn2

• • • fnn

 

Then provided the first-order conditions hold at a point [x?,

 

, x°„] the second-

order sufficient conditions for f (x?) to be a relative constrained maximum are:

( — 1)k Ifik I > 0, k = 2, ... , n,

 

 

 

 

(A.54)

whilst for the conditions for a relative constrained minimum are:

Ifik <

k = 2, ... , n.

(A.55)

If there are multiple constraints then additional Lagrange multipliers are added to the Lagrangian (one per constraint) and the first-order condition for each Lagrange

671

Mathematical Appendix

multiplier, Al , takes the form LAi aLiaa.; = 0. See Chiang (1984, pp. 385-386) for the appropriately defined bordered Hessian for the multi-constraint case.

Interpretation of the Lagrange multiplier

We now return to the single constraint case in order to demonstrate the interpretation of the Lagrange multiplier in the optimum. Using the superscript "0" to denote optimized values, we can write the optimized value of the Lagrangian as:

LO -C I 0 0

,

0‘ , 0

, ..,x)].

(A.56)

X2

, Xn ) + A [C. g(x7, x2()

Next, we ask the question what happens if the constraint is changed marginally. Obviously, both A,13 and x? are expected to change if c does. Differentiating (A.56) we get:

dr°

n

dx9

d)°

xo (dc) = )1/4,0,

(A.57)

dc

Li

1

) +

dc

dc

i=1

dc

 

 

 

 

 

 

 

 

where we have used the necessary conditions for an optimum (CA = Li = 0 for i = 1, 2, , n) to get from the first to the second equality. Recall that the constraint holds with equality (c = g(.)) so that A° measures the effect of a small change in c on the optimized value of the objective function f (.). For example, if the objective function is utility and c is income, then A° is the marginal utility of income.

A.4.3 Inequality constraints

We now briefly study some key results from non-linear programming. We first look at the simplest case with non-negativity constraints on the choice variables. Then we take up the more challenging case of general inequalities. We focus on first-order conditions and ignore some of the subtleties involved (like constraint qualifications and second-order conditions).

Non- negativity constraints

Suppose that the issue is to maximize a function y = f (x) subject only to the nonnegativity constraint x > 0. There are three situations which can arise. These have been illustrated in Figure A.1 which is taken from Chiang (1984, p. 723).

Panel (a) shows the case we have studied in detail above. The function attains a maximum for a strictly positive value of x. We call this an interior solution because the solution lies entirely within the feasible region (and not on a boundary). The constraint x > 0 is non-binding and the first-order condition is as before:

f'(xo) = 0. (interior solution)

Panels (b) and (c) deal with two types of boundary solutions. In panel (b) the function happens to attain a maximum for x = xo = 0, i.e. exactly on the boundary of the

feasible region. In panel

f'(xo) =-- 0 and I

Finally, in panel (c) we al f (x) continues to rise have:

f' (xo) 0 and

I These three conditions,

solutions, can be combi I

f' (x0) 0, xo

There are two key things A.1, we can safely ex, _

672