back start next


[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [ 71 ] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]


71

We will answer these questions in the following sections. But first, we will consider an example to illustrate the problem.

Illustrative Example

Table 5.1 presents consumption expenditures (y) and income (x) for 20 families. Suppose that we estimate the equation by ordinary least squares. We get

= 0.847 + 0.899X

(0.703) (0.0253)

R- = 0.986 RSS = 31.074

(Figures in parentheses are standard errors.) We can compute the predicted values and the residuals. In Table 5.2 we present the residuals for the observations which are ordered by their x-values. As we can easily see they are larger (absolutely) for larger values of x. Thus there is some evidence that the error variances are not constant but increase with the value of x. In Figure 5.1 we show the plot of the residuals. This shows graphically (perhaps more than Table 5.2) that there is a heteroskedasticity problem.

Sometimes, the heteroskedasticity problem is solved by estimating the regression in a log-linear form. When we regress log on log x, the estimated equation is

log = 0.0757 + 0.9562 log x

(0.0574) (0.0183)

R = 0.9935 RSS = 0.03757

(Figures in parentheses are standard errors.) The i?-s are not comparable since the variance of the dependent variable is different. We discuss the problem of comparing Rs from linear versus log-linear form in Section 5.6. The residuals

Residual +2

-2 -

• 10

Figure 5.1. Example of heteroskedasticity.



Table 5.1 Consumption Expenditures (y) and Income (jc) for 20 Families (Thousands of Dollars)

Family

Family

19.9

22.3

31.2

32.3

33.1

34.5

31.8

36.6

33.5

38.0

12.1

12.1

13.1

14.1

40.7

42.3

14.8

16.4

21.6

24.1

38.6

44.7

29.3

30.1

25.5

26.1

25.0

28.3

10.3

10.3

17.9

18.2

38.8

40.2

19.8

20.1

Table 5.2 Residuals for the Consumption Function Estimated from Data in Table 5.1 (Ordered by Their jc-Values)"

Observation

Value ofx

Residual

Observation

Value ofx

Residual

-0.32

26.1

1.18

-0.13

28.3

-1.30

10.3

0.19

30.1

1.38

12.1

0.37

32.3

1.30

14.1

-0.43

34.5

1.23

16.4

-0.80

36.6

-1.96

18.2

0.69

38.0

-1.52

20.1

0.88

40.2

1.80

22.3

-1.00

42.3

1.81

24.1

-0.92

44.7

-2.45

"Residuals are rounded to two decimals.

from this equation are presented in Table 5.3. In this situation there is no perceptible increase in the magnitudes of the residuals as the value of : increases. Thus there does not appear to be a heteroskedasticity problem.

In any case, we need formal tests of the hypothesis of homoskedasticity. These are discussed in the following sections.

5.2 Detection of Heteroskedasticity

In the illustrative example in Section 5.1 we plotted the estimated residual , against x, to see whether we notice any systematic pattern in the residuals that suggests heteroskedasticity in the errors. Note, however, that by virtue of the



Observation

logx

Residual

Observation

logx

Residual

1.82

-0.12

3.26

0.44

2.09

0.04

3.34

-0.53

2.33

0.27

3.40

0.47

2.49

0.34

3.48

0.42

2.65

-0.33

3.54

0.38

2.80

-0.56

3.60

-0.59

2.90

0.35

3.64

-0.42

3.00

0.41

3.69

0.51

3.10

-0.54

3.74

0.50

3.18

-0.46

3.80

-0.56

"Log x rounded to two decimals and residual multiplied by 10.

normal equation, , and x, are uncorrelated though ? could be correlated with X,. Thus if we are using a regression procedure to test for heteroskedasticity, we should use a regression of u, on x;, xJ, ... or a regression of , or on x„ X?, X?, .... In the case of multiple regression, we should use powers of S, the predicted value of y„ or powers of all the explanatory variables.

1. The test suggested by Anscombe and a test called RESET suggested by Ramsey both involve regressing , on y?, y?, . . . and testing whether or not the coefficients are significant.

2. The test suggested by White involves regressing uJ on all the explanatory variables and their squares and cross products. For instance, with three explanatory variables x,, , Xj, it involves regressing on x,, X2, Xj, X], xi, X3, XiXi, x,x„ and X3X,.

3. Glejser suggested estimating regressions of the type , = a + (3x„ , = a -I- p/Xj, , = a -I- Vx„ and so on, and testing the hypothesis

p = o.

The implicit assumption behind all these tests is that var(M,) = of = aiz) where z, is an unknown variable and the different tests use different proxies or surrogates for the unknown function /( ).

F. J. Anscombe, "Examination of Residuals," Proceedings of the Fourth Berlceley Symposium on Mathematical Statistics and Probability (Berkeley, Calif.: University of California Press, 1961), pp. 1-36.

J. B, Ramsey, "Tests for Specification Errors in Classical Linear Least Squares Regression Analysis," Journal of the Royal Statistical Society. Series B, Vol. 31, 1969, pp. 350-371. H. White, "A Heteroskedasticity Consistent Covariance Matrix Estimator and a Direct Test of Heteroskedasticity," Econometrica. Vol. 48, 1980, pp. 817-838.

H. Glejser, "A New Test for Homoscedasticity," Journal of the American Statistical Association. 1969, pp. 316-323.



[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [ 71 ] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]