back start next
[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [ 32 ] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]
32 3.6 Analysis of Variance for the Simple Regression Model Yet another item that is often presented in connection with the simple linear regression model is the analysis of variance. This is the breakdown of the total sum of squares TSS into the explained sum of squares ESS and the residual sum of squares RSS. The purpose of presenting the table is to test the significance of the explained sum of squares. In this case this amounts to testing the significance of p. Table 3.3 presents the breakdown of the total sum of squares. Under the assumptions of the model, RSS/cr has a x-distribution with (n -2) degrees of freedom. ESS/a, on the other hand has a x-distribution with 1 degree of freedom only if the true p is equal to zero. Further, these two x distributions are independent. (All these results are proved in the appendix to this chapter.) Thus under the assumption that = 0, we have the F-statistic (cr cancels) F = (ESS/l)/RSS/(/j - 2), which has an F-distribution with degrees of freedom 1 and (n - 2). This F-statistic can be used to test the hypothesis that p = 0. For the data in Table 3.2 we are considering, the analysis of variance is presented in Table 3.4. The F-statistic is F = 15.75/1.83 = 8.6. Note that the /-statistic for testing the significance of p is p/SE(P) = 0.75/0.256 = 2.93 and the F-statistic obtained from the analysis of variance Table 3.4 is the square of the /-statistic. Thus in the case of a simple linear regression model, the analysis-of-variance table does not give any more information. We will see in Chapter 4 that in the case of multiple regression this is not the case. What the analysis-of-variance table provides there is a test of significance for all the regression parameters together. Note that Table 3.3 Analysis of Variance for the Simple Regression Model Source of | | Degrees of | | Variation | Sum of Squares | Freedom | Mean Square | | ESS = p5 | | ESS/1 | Residual | RSS = 5 - PS | | RSS/(/j - 2) | Total | TSS = s,. | n - 1 | |
Table 3.4 Analysis of Variance for the Data in Table 3.2 Source of | Sum of | Degrees of | Mean | Variation | Squares | Freedom | Square | | 15.75 | | 15.75 | Residual | 14.65 | | 1.83 | Total | 30.4 | | |
t + (n - 2) We can check in our illustrative example that = 8.6 and (n - 2) = 8. Hence 8.6 8.6 as we obtained earlier. The formula t + in- 2) gives the relationship between the /-ratio for testing the hypothesis p = 0 and the r. We will derive a similar formula in Chapter 4. 3.7 Prediction with the Simple Regression Model The estimated regression equation = d + px is used for predicting the value of for given values of x and the estimated equation x = d + py is used for predicting the values of x for given values of y. We will illustrate the procedures with reference to the prediction of given x. Let Xo be the given value of x. Then we predict the corresponding value of of by = d + pxo (3.11) The true value y is given by = a + pxo + Mo where Ug is the error term. Hence the prediction error is = (a - a} + - p)xo - «0 ESS = 05„ = rS and RSS = S„ - pS„ = (1 - ,)S,, Hence the F-statistic can also be written as ( -2 (I - 1}/{ - 2) 1 - SSnce F = 1 we get
Since £(a - a) = 0, - ) = 0, and £(t<o) = 0 we have £(50 - ) = 0 This equation shows that the predictor given by equation (3.11) is unbiased. Note that the predictor is unbiased in the sense that £( ) = £( ) since both So and are random variables. The variance of the prediction error is a, Vifo - ) = Via - a) + xVi - p) + 2xo cov(a , /1 x ) + V(mo) + cjf-2xocrf + a 1 + - + -- Thus the variance increases the farther away the value of Xq is from x, the mean of the observations on the basis of which a and p have been computed. This is illustrated in Figure 3.5, which shows the confidence bands for y. If Xo lies within the range of the sample observations on x, we can call it within-sample prediction, and if Xq lies outside the range of the sample observations, we call the prediction out-of-sample prediction. As an illustration, consider a consumption function estimated on the basis of 12 observations. The equation is (this is just a hypothetical example) Figure 3.5. Confidence bands for prediction of given jc. [Confidence bands shown are ± Z, where Z = tSE(y) and / is the f-value used.]
[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [ 32 ] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]
|