back start next


[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [ 83 ] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]


83

Since Po = 1 we get by successive substitution

Pi = P. = P = p. • •

Thus the lag correlations are all powers of p and decline geometrically.

These expressions can be used to derive the covariance matrix of the errors and using what is known as generalized least squares (GLS). We will not derive the expression for GLS here but will outline the minor changes that it implies. Consider the model

y, = a + x, + u, t=l,2,...,T (6.2)

M, = pM, + e, (6.3)

Except for the treatment of the first observation in the case of AR(1) errors as in (6.3), and the treatment of the first two observations in the case of AR(2) errors, and so on, the GLS procedure amounts to the use of transformed data, which are obtained as follows.*

Lagging (6.2) by one period and multiplying it by p, we get

py, , = ap + (3pj:, , + , , (6.4)

Subtracting (6.4) from (6.2) and using (6.3), we get

y, - /-1 = a(l - P) + - px, i) + e, (6.5)

Since e, are serially independent with a constant variance ct?, we can estimate the parameters in this equation by an OLS procedure. Equation (6.5) is often called the quasi-difference transformation of (6.4). What we do is transform the variables y, and x, to

y* = y, - py,-i / = 2, 3, . . . ,

x = X, - px,.,

and run a regression of y* on x*, with or without a constant term depending on whether the original equation has a constant term or not. In this method we use only (J-l) observations because we lose one observation in the process of taking differences. This procedure is not exactly the GLS procedure. The GLS procedure amounts to using all the observations with

y\ = VI - pyi (6.6)

x\ = Vl - px,

and regressing y* on jc* using the T observations.

In actual practice p is not known. There are two types of procedures for estimating p:

1. Iterative procedures.

2. Grid-search procedures,

The derivation involves the use of matrix algebra, which we have avoided. 4f the number of observations is large, the omission of these initial observations does not matter much.



Iterative Procedures

Among the iterative procedures, the eariiest was the Cochrane-Orcutt procedure. In the Cochrane-Orcutt procedure we estimate equation (6.2) by OLS, get the estimated residuals „ and estimate p by p = 2]

Durbin" suggested an alternative method of estimating p. In this procedure, we write equation (6.5) as

y, = a(l - p) + py,„, + Px, - Ppx, , + e, (6.7)

We regress y, on y, , x„ and x, , and take the estimated coefficient of y,, as an estimate of p.

Once an estimate of p is obtained, we construct the transformed variables y* and X* as defined in (6.6) and (6.6) and estimate a regression of y* on x*. The only thing to note is that the slope coefficient in this equation is p, but the intercept is a(l - p). Thus after estimating the regression of y* on x*, we have to adjust the constant term appropriately to get estimates of the parameters of the original equation (6.2). Further, the standard errors we compute from the regression of y* on x* are now "asymptotic" standard errors because of the fact that p has been estimated. If there are lagged values of as explanatory variables, these standard errors are not correct even asymptotically. The adjustment needed in this case is discussed in Section 6.7.

If there are many explanatory variables in the equation, Durbins method involves a regression in too many variables (twice the number of explanatory variables plus y,„). Hence it is often customary to prefer the Cochrane-Orcutt procedure until it converges. However, there are examples" to show that the minimization of 2] e; in (6.5) can produce multiple solutions for p. In this case the Cochrane-Orcutt procedure, which relies on a single solution for p, might give a local minimum, and even when iterated might converge to a local minimum. Hence it is better to use a grid-search procedure, which we will now describe.

Grid-Search Procedures

One of the first grid-search procedures is the Hildreth and Lu procedure suggested in 1960. The procedure is as follows. Calculate y* and x* in (6.6) for different values of p at intervals of 0.1 in the range - 1 < p < 1. Estimate the regression of y* on xj and calculate the residual sum of squares RSS in each

D. Coctirane and G. H. Orcutt, "Application of Least Squares Regressions to Relationships Containing Autocorrelated Error Terms," Journal of the American Statistical Association, 1949, pp. 32-61.

"J. Durbin, "Estimation of Parameters in Time Series Regression Models," Journal of the Royal Statistical Society, Series B. 1960, pp. 139-153.

"J. M. Dufour, M. J. 1. Gaudry, and T. C. Lieu, "The Cochrane-Orcutt Procedure: Numerical Examples of Multiple Admissible Minima," Economics Letters, 1980 (6), pp. 43-48. -Clifford Hildreth and John Y. Lu, Demand Relations with Autocorrelated Disturbances, AES Technical Bulletin 276, Michigan State University, East Lansing, Mich., November 1960.



case. Choose the value of p for which the RSS is minimum. Again repeat this procedure for smaller intervals of p around this value. For instance, if the value of p for which RSS is minimum is - 0.4, repeat this search procedure for values of p at intervals of 0.01 in the range -0.5 < p < -0.3.

This procedure is not the same as the maximum likelihood (ML) procedure. If the errors e, are normally distributed, we can write the log-likelihood function as (derivation is omitted)

where

T 1 Q

log L = const. - - log CT? -I- - log (1 - P) -

e = S [y, - py, , - a(l - p) - ?>ix, - px, ,)]2

(6.8)

Thus minimizing Q is not the same as maximizing log L. We can use the grid-search procedure to get the ML estimates. The only difference is that after we compute the residual sum of squares RSS(p) for each p, we choose the value of p for which {TIT) log RSS(p) - (1/2) log (1 - p) is minimum. If the number of observations is large, the latter term will be small compared to the former, and the ML procedure and the Hildreth-Lu procedure will give almost the same results.

Illustrative Example

Consider the data in Table 3.11 and the estimation of the production function log Z = a -I- (3, log L, + p, log , + M

The OLS estimation gave a DW statistic of 0.86, suggesting significant positive autocorrelation. Assuming that the errors were AR(1), two stimation procedures were used: the Hildreth-Lu grid search and the iterative Cochrane-Orcutt (C-O). The other procedures we have described can also be tried, but this is left as an exercise.

The Hildreth-Lu procedure gave p = 0.77. The iterative C-O procedure gave p = 0.80. The DW test statistic implies that p = (l/2)(2 - 0.86) = 0.57.

The estimates of the parameters (with the standard errors in parentheses) were as follows:

Estimate of

Hildreth-Lu

Iterative C-O

-3.938

-2.909

-2.737

(0 237)

(0.4*2)

(0 461)

1.451

1.092

1.070

(0 083)

(0 151)

(0 153)

0.384

0.570

0.558

(0 048)

(0 104)

(0 097)

0.04338

0.02635

0.02644

In this example the parameter estimates given by Hildreth-Lu and the iterative C-O procedures are pretty close to each other. Correcting for the auto-



[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [ 83 ] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]