Добавил:
Upload Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:

EH_ch1_4

.pdf
Скачиваний:
8
Добавлен:
26.03.2016
Размер:
549.06 Кб
Скачать

Chapter 3

Variations on a Theme

3.1 Introduction

In this chapter we discuss some extensions and variations of econometric learning. Several issues arise naturally. So far we have assumed representative agent learning, although diversity of expectations should surely be treated. One can also consider alternative adaptive learning schemes and the possibility that the agents do not know the true model. In this chapter we show how such issues can be readily addressed in the context of the basic cobweb and asset pricing models.

We also take up learning in nonstochastic frameworks and obtain the key conditions for local stability under adaptive learning of perfect-foresight steady states. Since some standard textbook models are often presented in a nonstochastic setting, we will occasionally draw on these results in later chapters. In this chapter we show how the technique can be applied directly to assess local stability of equilibria in simple nonstochastic coordination games.

3.2 Heterogeneous Expectations

For expository purposes we simplify the cobweb model of the previous chapter. Dropping the intercept and assuming a scalar stationary shock wt1, the model

45

46

 

 

 

 

 

View of the Landscape

becomes1

 

 

 

 

 

 

 

pt = αpte + δwt1 + ηt .

 

(3.1)

Recall that the RE solution is given by

 

 

 

 

 

t = ¯

t1 + t

¯

=

(1

α)1δ.

p bw

η ,

where b

 

 

We allow for i = 1, . . . , N different groups of agents who may have different

expectations pe

, but assume that the average expectations

i,t

 

 

 

 

N

 

pte = N1

 

 

pi,te

 

 

i=1

influence the market price in equation (3.1). Each group of agents forecasts according to the linear rule

pi,te = bi,t1wt1.

Thus the agents are forecasting in the same way, but they are allowed to have different parameter estimates.

We continue to assume that agents learn from the data on past prices and the exogenous variables and use a variation of least squares learning to update their estimates. In fact, we allow for a slight generalization of recursive least squares as follows:

b

i,t =

b

i,t1

+

γ R1w (p

b

i,t1

w

),

 

 

t

i,t

t1 t

 

t1

(3.2)

Ri,t =

Ri,t1 + γt

wt21 Ri,t1

.

 

 

In the previous chapter the least squares formula set γt = t1. We have now introduced a more general gain parameter γt , as discussed in Section 2.7 of that chapter.

Combining the earlier equations for the individual expectations with equation (3.1) leads to

pt = N

1α iN1 bi,t1

wt1 + δwt1 + ηt .

 

 

 

 

=

 

1The argument can be readily generalized to allow for an intercept and a vector stationary shock wt . In fact, one can also allow for pt to be a vector. Details are provided in Evans and Honkapohja (1997).

Variations on a Theme

47

This equation and the updating rule (3.2) define a standard stochastic recursive algorithm as introduced in Section 2.7. As pointed out there, the analysis of convergence in these algorithms is carried out by deriving the associated ODE. Following the same steps as in Section 2.8 of Chapter 2, we obtain

 

dbi

= Ri1M δ + N1α

N

, i = 1, . . . , N,

 

 

i=1 bi bi

 

dRi

= M Ri ,

 

i = 1, . . . , N,

 

 

where M = E(wt2).

Since Ri M globally for i = 1, . . . , N, stability of this differential equation system is governed by stability of

 

N

dbi

 

= δ + N1α bi bi , i = 1, . . . , N.

 

i=1

This system can be written in the matrix form

 

.

 

N

 

α

. .

db

 

δ

 

 

 

 

 

 

1 . . .

=

.

 

+

 

1

 

 

.

δ

 

1 . . .

 

 

.

 

 

 

 

 

 

. . .

 

 

 

 

 

 

1

.

.. IN b.

1

The coefficient matrix of this linear system has one eigenvalue of α 1 and N 1 eigenvalues of 1 and is therefore globally asymptotically stable with

each bi converging to

¯

provided

α <

1.

 

b,

 

 

Appealing to the results on convergence of stochastic recursive algorithms as in the previous chapter, we have the striking result that the convergence condition under heterogeneous expectations is the same as that for the case of a representative agent. That is, provided α < 1, all agents learn asymptotically the rational expectations equilibrium.

We remark that the preceding argument implicitly makes the “representative agent” assumption of structural homogeneity. Thus equation (3.1) is assumed to arise from an economy in which the cost functions of individual suppliers are identical. In principle structural heterogeneity could be tackled by the same techniques.

48

View of the Landscape

3.3 Learning with Constant Gain

So far we have considered cases where there is asymptotic convergence to the rational expectations equilibrium. In models with stochastic shocks, this can only happen if the learning algorithm is of the “decreasing-gain” type, i.e., γt 0. We now briefly analyze the case of a constant gain, i.e., γt = γ , a small positive constant.

For our example we pick the model

pt = α + βEt pt+1 + vt ,

where vt is iid with mean zero and Et pt+1 denotes the (rational or nonrational) expectation of the next period price. We here use Et pt+1 in place of pte+1 to emphasize that the expectations are formed at time t. This model was introduced in Chapter 1 as the “Cagan Model” and it was noted there that it could also be interpreted as the standard asset pricing model with risk neutrality.2 For both cases we have 0 < β < 1.

We focus on learning of the market fundamental rational expectations solution. This is a stochastic steady state pt = a¯ + vt , where a¯ = (1 β)1α. We first begin with the standard decreasing-gain learning rule which does converge to the REE. We assume that forecasts take the form Et pt+1 = at , where at is the estimated mean which is updated according to

at = at1 + t1(pt1 at1).

Since pt1 = α + βat1 + vt1, we have

at = at1 + t1+ 1)at1 + vt1).

The associated ODE takes the form

da = α + 1)a. dτ

Applying the stochastic approximation results, it follows that at a¯ provided β < 1. (Using the global convergence results of Chapter 6, it can in fact be shown that convergence occurs with probability 1.) The condition β < 1 can easily be seen to be the E-stability condition, so that we obtain the expected result that when this condition is satisfied there is convergence to the REE.

2For further discussion of this model see Section 8.6 of Chapter 8.

Variations on a Theme

49

Now we consider the implications of replacing t1 by a constant parameter γ . The forecasts still take the form Et pt+1 = at , but now at is updated using the constant-gain learning rule

at = at1 + γ (pt1 at1), where 0 < γ 1.

(3.3)

We remark that this learning rule is equivalent to the traditional adaptive expectations formula. Note also that it can be expressed as an exponentially weighted

average of lagged prices since a

t =

γ

(1

γ )i p

.

 

 

 

i=0

 

t1i

 

 

Since pt = α + βat + vt , we

have

 

 

 

 

 

 

 

 

 

 

 

 

 

 

at = αγ + 1 γ (1 β) at1 + γ vt1.

(3.4)

This is an AR(1) process which is stationary if |1 γ (1 β)| < 1 or equivalently, 2 > γ (1 β) > 0. A necessary condition for stationarity is that β < 1, which is also the weak E-stability condition for this model. In the limit γ 0, the E-stability condition β < 1 is also sufficient for stationarity.

The price process takes the form

pt = 1 γ (1 β) pt1 + αγ + vt (1 γ )vt1.

This is an ARMA(1,1) process. Assuming |1 γ (1 β)| < 1 so that both at and pt are (asymptotically) stationary, it is easily verified that the (asymptotic) mean of both pt and at are equal to the RE value a¯ = (1 β)1α. Thus the forecast at is asymptotically unbiased. It is also possible to compute the (asymptotic) variance of pt .3 This is given by

var(p )

=

1 + (1 γ )(1 2β)

 

var(v ).

1 + (1 γ )(1 2β) γβ2

t

t

For γ > 0 the variance is higher than the RE value of var(vt ), though it approaches this as γ 0. This illustrates the phenomenon of excess volatility induced by the fixed-gain learning rule.

3For a stationary ARMA(1,1) process

yt = φyt1 + εt + θεt1,

where |φ| < 1 and εt is white noise with variance σ 2, it can be shown that

var(yt ) = (1 + θ2 + 2φθ) σ 2. 1 φ2

See, for example, Harvey (1981).

50

View of the Landscape

Although, as just demonstrated, fixed-gain learning rules do not generally converge to rational expectations, they can do so if the model is nonstochastic. Thus suppose that vt 0, so that

pt = α + βEt pt+1.

With forecasts Et pt+1 = at and learning rule (3.3), we have convergence to the perfect-foresight steady state pt = α/(1 β) if and only if |1 γ (1 β)| < 1, i.e.,

1 2/γ < β < 1.

Note that if |β| < 1, this holds for all 0 < γ 1, whereas for β < 1 we need γ < 2(1 β)1. Thus, if the expectational stability condition β < 1 holds, in the nonstochastic model there is always convergence to the RE under the constantgain learning rule if the gain parameter γ is sufficiently small.

3.4 Learning in Nonstochastic Models

We now take up the issue of learning steady states in nonstochastic models in a more general setting. Consider models of the form

pt = f (Et pt+1).

One natural adaptive learning rule is to forecast pt+1 as the average of past observed values, i.e.,

Et pt+1 = at ,

where at = t1 t1 pi , for t = 1, 2, 3, . . . . This can be written recursively as

i=0

at = at1 + t1(pt1 at1).

As discussed for a linear model in the previous section, in nonstochastic models the constant-gain version also has the potential to converge to a perfect-foresight steady state. This remains true for local analysis of nonlinear models. Both of these cases are covered by the following recursive formulation:

at = at1 + γt (pt1 at1),

(3.5)

Variations on a Theme

51

where the gain sequence γt satisfies

 

 

0 < γt 1 and

 

γt = +∞.

 

t=1

The choice γt = t1 is an example of a “decreasing-gain” algorithm (i.e., an algorithm in which limt→∞ γt = 0).

The formulation (3.5) reflects the common practice of assuming that the parameter estimate at depends only on data through t 1. This has the advantage of avoiding simultaneity between pt and at . However, we also briefly consider

the implications of an alternative assumption in which

 

at = at1 + γt (pt at1).

(3.6)

We will see that the choice of assumptions can matter in the fixed-gain case, but is not important if γt = t1 or if γt = γ > 0 is sufficiently small.

Standard timing assumption: We consider first the case (3.5), which we will refer to as the standard timing assumption. Combining equations, we have pt = f (at ) so that

at = at1 + γt (f (at1) at1).

In a perfect-foresight steady state, pt = p¯ = a, where a = f (a). In the constantgain case, we have at = (1 γ )at1 + γf (at1). To analyze stability we apply standard results on nonlinear difference equations.4 Using these, it is

easily established that a steady

state a = p¯ is locally stable

if and

only if

|

1

+

γ (f (a)

|

2/γ < f (a) < 1. Note that 1

2

→ −∞

 

 

 

1) < 1, i.e., iff 1

 

 

 

as γ 0.

 

 

 

 

 

 

 

 

 

 

Under the decreasing-gain assumption limt→∞ γt = 0, it can be shown that

a steady state a

= ¯

 

 

 

 

 

p is locally stable if and only if f (a) < 1.5 We here present

a proof of this for the case 0 < f (a) < 1. [For f (a) < 0, the argument is more involved.] Without loss of generality we suppose the initial point a0 < a. We assume that a0 is sufficiently close to a so that f is monotonically increasing

4See Chapter 5 for a review of the stability results. The key result is that a steady state y¯ of a multivariate difference equation yt = F (yt1) is locally stable if the derivative matrix DF (y)¯ has all eigenvalues inside the unit circle. In the case at hand we have a univariate difference equation, and so the condition is just that the derivative is less than 1 in absolute value.

5The results for the decreasing-gain case in this section can be shown formally by using the implicit function theorem and applying Evans and Honkapohja (2000). Details are given in Chapter 7.

52 View of the Landscape

between a0 and a. Then at each time t one has f (at ) > at , implying at+1 > at . Since γt < 1, it follows that at+1 a. To show that at a, suppose to the

contrary that lim at = aˆ < a.6 Let d

= f (a)ˆ − aˆ > 0. For t sufficiently large,

a

t+1

a

t +

γ

(d/2), so that a

t+s

a

t +

(d/2)

t+s

γ

, yielding a contradiction

 

 

t

= ∞

 

 

i t

i

 

 

 

 

=

 

 

 

 

 

 

=

 

 

since

t

1

γt

.

 

 

 

 

 

 

Alternative timing assumption: Under equation (3.6) we instead have the implicit equation

at = at1 + γt (f (at ) at1).

In general, there need not be a unique solution for at given at1, though this will be assured if γt is sufficiently small and if f (a) is bounded. Assuming uniqueness and focusing on local stability near a, we can approximate this equation by

at a = 1 γt f (a) 1(1 γt )(at1 a).

Under the fixed-gain assumption this leads to the stability condition that either f (a) < 1 or f (a) > 21 (these possibilities correspond to the cases γf (a) < 1 and γf (a) > 1). Under decreasing gain the condition is again simply f (a) < 1.

The results of this section can be summarized as follows. Fixed-gain learning can converge to perfect-foresight steady states in nonstochastic models and such rules have somewhat different stability conditions than the decreasing-gain rules which are standard for stochastic models. The stability conditions also depend on assumptions made concerning the timing of information. However, for the small gain case, i.e., decreasing gain or a sufficiently small constant gain, the condition for local stability under adaptive learning is not affected by the timing assumption and is simply f (a) < 1, generalizing our earlier results for linear models. This stability condition has a straightforward interpretation in terms of E-stability: For a PLM yt = a, the corresponding ALM is yt = f (a). The E-stability differential equation is then da/dτ = f (a) a with stability condition f (a) < 1.

The methods of this section can also be easily applied equally to models of the form pt = f (Et1pt ). For this setup, the expectation Et1pt cannot depend on pt and the timing assumption appears clear. Supposing that Et1pt = at1 with at = at1 + γt (pt at1) leads to the recursive algorithm

6Recall that a bounded monotonic sequence always has a limit.

Variations on a Theme

53

at = at1 + γt (f (at1) at1) as with the preceding model under the standard timing assumption. It follows that a perfect-foresight steady state pt = p¯, where p¯ = f (p)¯ , is locally stable under decreasing gain if f (a) < 1 and under constant gain γt = γ provided 1 2/γ < f (a) < 1.

3.4.1Application: Coordination Problems

The general methods of this section can be applied to study adaptive learning in certain nonstochastic macroeconomic models formulated as games. In this section we consider coordination games involving strategic complementarities along the lines of Cooper and John (1988) and Cooper (1999). A large finite number of agents i = 1, . . . , I, each choose an action xi [0, 1]. We consider a representative agent model in which the payoff for agent i is given by U (xi , Y (xi )), where xi denotes the vector of actions by other agents and Y (xi ) is an aggregate statistic. U (xi , Y (xi )) is assumed to be twice continuously differentiable and strictly concave in the first argument. We focus on symmetric outcomes in which each agent chooses action x and we adopt the notation Y (x) for Y (xi ). We assume that Y (x) is twice continuously differentiable and that Y (x) > 0 for all x [0, 1].

A symmetric Nash equilibrium is an action x which maximizes U (x, Y ) with respect to x, given Y , and where Y = Y (x). In general, there may be multiple equilibria. For this game, a necessary condition for this is strategic complementarity, defined by 2U/∂x∂Y > 0. Let φ(Y ) denote the best response of the representative agent to Y . In the case of strategic complementarity we obtain an increasing function φ(Y (x)) as in Figure 3.1.

Figure 3.1.

54 View of the Landscape

We illustrate the common case of three steady states, given by the fixed points x = φ(Y (x)). Obviously, the number of equilibria could be larger or smaller than three in the case of strategic complementarity. Note that x¯ = φ(Y (x))¯ is a fixed point of φ(Y (x)) if and only if a¯ = Y (x)¯ is a fixed point of Y (φ(a)).

The above description is presented as a static game, but is often used in the context of repeated games. We continue to focus on the static Nash equilibria. Agents now have available information on past outcomes which they can use to forecast the behavior of other agents and in particular the value of the key aggregate statistic Y .

Let xt be the action of the representative agent at time t and Yt be the value of Y (xt ) at t. Stability under learning can be treated following the general techniques developed in this section. At each time t, agents forecast the aggregate statistic Y and choose the optimal action conditioned on this forecast. Letting at denote the expectation of Yt , we assume

at = at1 + γt (Yt1 at1),

where γt is the gain sequence. Here we use the standard timing assumption.7 Given this expectation for Yt , the representative agent chooses action xt = φ(at ) so that

Yt = Y φ(at ) .

From the general analysis of this section we immediately conclude that a fixed point a¯ of Y (φ(a)) is locally stable under this learning rule for at if and only

if dY (φ(a))/da¯

< 1. It follows8 that an equilibrium x¯ is locally stable if and

only if dφ(Y (x))/dx¯

< 1. In Figure 3.1, the lowest and highest steady states

are thus stable, while the middle one is unstable. Since dφ(Y (x))/dx¯

> 0 in

the case of strategic complementarities, this stability condition holds both for decreasing-gain sequences and also for all constant gains 0 < γ 1.

There are many economic settings that fit into the framework of strategic complementarities. The underlying economic mechanisms rely on diverse phenomena such as technological complementarities, imperfect competition, demand spillovers, and search externalities.9 Cooper (1999) gives an up-to-date

7The alternative timing assumption of contemporaneous information seems particularly unnatural in this model.

8Use dY (φ(a))/da¯ = Y (φ(a))φ¯ (a)¯ = Y (x)φ¯ (Y (x))¯ .

9Early papers in the extensive literature on coordination problems include Bryant (1983), Bryant (1987), Diamond (1982), Hart (1982), Schleifer (1986), and Weitzman (1982).

Соседние файлы в предмете [НЕСОРТИРОВАННОЕ]