back start next


[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [ 164 ] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]


164

1 for the (n + l)th observation 0 for all others

We have discussed this procedure in Chapter 8.

Now we just run a multiple regression with all the (« + 1) observations and this extra variable D. The estimate of the coefficient of this variable D is the prediction error and its standard error is ad„+f. This result has been proved in Chapter 8. Thus we get from the regression program and d„+i a or t„+i the /-ratio, which is e„+y/d„+, a. But for the recursive residuals we need e„+/J„+,. Thus all we have to do is multiply the /-ratio we get for the coefficient of D by . This gives us the recursive residual.

The calculation of recursive residuals by the dummy variable method is similar to that for predicted residuals except that the regressions for predicted residuals involve all the observations in the sample and for recursive residuals the observation set is sequential and the dummy variable is also sequential.

Illustrative Example

In Table 12.1 we present the OLS residual, the forward recursive, and the backward recursive residuals for the production function (4.24). The sum of squares of the recursive residuals should be equal to the sum of squares of the OLS residuals, but there are small discrepancies which are due to rounding errors. The recursive residuals are useful for:

1. Tests for heteroskedasticity described in Chapter 5.

2. Tests for autocorrelation described in Chapter 6.

3. Tests for stabihty described in Chapter 4.

For stability analysis. Brown, Durbin, and Evans suggest computing the cumulative sums (CUSUM) and cumulative sums of squares (CUSUMSQ) of the recursive residuals and comparing them with some percentage points that they have tabulated. A discussion of this is beyond the scope of this book." Instead,

"Phillips and Harvey, "A Simple Test."

"Those interested in this detail can refer to the paper by Brown, Durbin, and Evans, "Techniques for Testing."

3. Their sum of squares is equal to RSS, the residual sum of squares from the least squares regression.

The third property is useful in checking the accuracy of the calculations.

There are algorithms for calculating the regression coefficients p, and also the prediction variance dfa, in a recursive fashion.* There is, however, an alternative method that we can use for the recursive residuals as we have done for the predicted residuals described eadier. Note that recursive residuals are similar to predicted residuals except that the predictions are sequential.

Suppose that we use n observations and want to get the prediction error and the variance of the prediction error for the (n + l)th observation. Then all we have to do is to create a dummy variable D which is defined as



Year

OLS Residual

Forward Recursive

Backward Recursive

1929

-1.27

-1.304

1930

-4.29

-4.530

1931

-3.99

-4.710

1932

-1.30

-3.044

1933

-3.04

-1.975

-6.187

1934

-1.99

-0.809

-6.476

1935

-0.79

0.512

-6.440

1936

1.04

1.115

-4.246

1937

-0.26

-1.040

-4.359

1938

4.55

3.824

-2.900

1939

4.93

4.163

-2.224

1940

4.37

4.182

-2.668

1941

1.84

1.335

-2.971

1942

-1.74

-1.456

-4.775

1943

0.13

1.609

-3.110

1944

6.59

6.065

3.502

1945

11.11

8.487

8.844

1946

1.34

-4.811

5.452

1947

-6.59

-10.322

-1.531

1948

-5.90

-7.172

-1.884

1949

-2.06

-1.601

-0.132

1950

0.12

1.343

2.331

1951

-2.68

-0.884

0.273

1952

-3.24

-0.667

-0.702

1953

-2.38

0.883

-0.716

1954

1.10

4.137

1.293

1955

1.48

3.540

2.917

1956

-1.52

0.186

0.826

1957

-1.00

0.772

0.764

1958

1.43

2.917

0.912

1959

0.00

1.083

0.989

1960

-1.39

-0.367

-0.854

1961

0.38

1.400

-2.372

1962

1.21

1.963

-1.994

1963

2.51

2.865

-1.793

1964

2.31

2.297

1965

1.85

1.561

1966

0.00

-0.328

1967

-2.97

-3.160

Mean

0.00

0.618

-1.252

3.379

3.506

3.338

Range

17.70

18.81

15.32

433.8

431.3

433.6

0.00

1.04

-2.22

Table 12.1 Different Residuals for the Production Function (4.25) (Multipled by 100)



12.5 DFFITS and Bounded Influence Estimation

In Section 3.8 we discussed briefly the problem of outliers. The usual approach to outliers based on least squares residuals is as follows:

1. Look at the OLS residuals.

2. Delete the observations with large residuals.

3. Reestimate the equation.

Two major problems with this approach are that the OLS residuals (as we showed earlier) do not all have the same variance and furthermore the OLS residuals do not give us any idea of how important this particular observation is for the overall results. The idea behind the studentized residual is to allow for these differences in the variances and to look at the prediction error resulting from the deletion of this observation. Using a plus or minus 2a rule of thumb, the studentized residuals shown in Table 12.2 suggest that the observations for the years 1945 and 1947 are outliers. With the OLS residuals in Table 12.1 we might have included 1944 and 1948 as well.

One other measure that is used to detect outliers is to see the change in the fitted value of that results from dropping a particular observation. Let y,,, be the fitted value of if the rth observation is dropped. The quantity (y, - y,,,) divided by the scaling factor h„S„ where Sj is the estimator of a from a regression with the ith observation omitted is called DFFITS,.

It has been shown that*

DFFITS, = I j- ) ,

where w, is the rth studentized residual and h,, is the quantity that figures repeatedly in equations (12.1)-(12.3). (It is also known as the rth diagonal term of the "hat matrix.")

There are many computer programs available to compute studentized residuals and DFFITS. For instance, the SAS regression program gives these statis-

"D. A. Belsley, E. Kuh, and R. E. Welsch, Regression Diagnostics: Identifying Influential Data and Sources of Collinearity (New York: Wiley, 1980).

we will apply some simple Mests. If a test of the hypothesis that the mean of the recursive residual is zero gives us a significant r-statistic, this is an indication of instability of the coefficients. In Table 12.1 we present these f-ratios. With the forward recursive residuals we do not reject the hypothesis of zero mean. With the backward recursive residuals, we do reject at the 5% level, the hypothesis of zero mean. Thus the conclusions from the recursive residuals are similar to the conclusions we arrived at eariier in Chapter 4 from the predictive tests for stabihty.



[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [ 164 ] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]