back start next


[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [ 54 ] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]


54

Thus, for the variable M,

The simple is just the square of the correlation of average cost per case and the proportion of the cases in that category. The two are presented together to illustrate how the partial rs can be lower or higher than the simple rs.

The regression coefficients in this example have to be interpreted in a different way. In the usual case, each regression coefficient measures the change in the dependent variable for unit increase in that independent variable, holding other variables constant. In this example, since the independent variables are proportions in that category; an increase in one variable holding other variables constant does not make sense (in fact, it is impossible to do it). The proper interpretation of the coefficients in this case is this: Put the value of M = 1, all others 0. Then the estimated value of the dependent variable = (constant + coefficient of M) = 69.51 + 44.97 = 114.48. This is the average cost of treating a case in M. Similarly, constant + coefficient of P = 69.51 - 44.54 = 24.97 is the average cost of treating a case in P. Finally, putting M, P, . . . , Ob all = 0, we get the constant term = 69.51 as the average cost of treating a case in the "Other" category. These coefficients are all presented in the last column of Table 4.2. What the regression equation enables us to estimate in this case is the average cost of treating a case in each category. The standard errors of these estimates can be calculated as SE(a + .) if we know the covariance between the constant term & and the other regression coefficient -

An important point to note is that we have not estimated a coefficient for the category "other." An alternative procedure in which we estimate coefficients for all the categories directly would be to include all the variables and delete the constant term. We have to delete the constant term because the sum of all the explanatory variables (which are proportions of cases treated in each category) is equal to 1 for all observations. Since the constant term corresponds to an explanatory variable that is identically equal to 1 for all observations, we cannot include it in the regression equation. If we do, we have a situation where one of the explanatory variables is identically the sum of the other explanatory variables. In this case we cannot estimate all the coefficients.

To see what the problem is, consider the equation

= a + P,x, + 22 + - + "

where = xi + xj. We can easily see that we cannot get separate estimates of 1, 2 and since the equation reduces to

= a + , :, + 22 + (1 + ) + = + ( , + ) 1 + ( + + "

such situation is known as "perfect multicollinearity" and is discussed in greater detail in Chapter 7.



Example 2: Demand for Food

Table 4.3 presents data on per capita food consumption, price of food, and per capita income for the years 1927-1941 and 1948-1962.* Two equations are estimated for these data:

Equation 1: log = a + , log p + P2 log

Equation 2: log = a -f 1 log P + 2 log + log p log

The last term \n Equation 2 is an interaction term that allows the price and income elasticities to vary. Equation 1 implies constant price and income elasticities.

*The data are from Frederick V. Waugh, Demand and Price Analysis: Some Examples from Agriculture, U.S.D.A. Technical BulleUn 1316, November 1964, p. 16. The value of j for 1955 has been corrected to 96.5 from 86.5 that table.

A regression equation with no constant term is also called "regression through the origin" because if the x-variables are all zero, the y-variable is also zero. The estimation of such an equation proceeds as before and the normal equations are defined as in equations (4.7) and (4.8) except that we do not apply the "mean corrections." That is, we define

511 = X xi and not X (xi, - xjf

512 = S Xi,X2, and not X (xi, - X)(x2, - Xj) etc.

After these changes are made, the solution of the normal equations proceeds as before. The definition of RSS and also proceeds as before, with the change that in 5 as well, no mean correction be applied.

Cautionary Note: Many regression programs allow the option of including or excluding the constant term. However, some of them may not give the correct R when running a regression equation with no constant term. If the R is computed as 1 - RSS/5„ and 5„ is computed using the mean correction, we might get a very small and sometimes even a negative R.

Note that with two explanatory variables, and no constant term, RSS = 2 >? - Pi(5] y.xu) - 02(2 . ) and if S„ is computed (wrongly) with a mean correction, so that 5„ = X ? ~ 3 happen that < RSS and thus we get negative R.

The correct R in this case is 1 - RSS/X ?- Of course, if is very high X y? will be very high compared with RSS and we might get a high R. But one should not infer from this that the equation with no constant term is better than the one with a constant term. The two R values are not comparable. With a constant term, the R measures the proportion of XCVi ~ explained by the explanatory variables. Without a constant term that R explains the proportion of 2 y explained by the explanatory variables. Thus we are talking of proportions of two different things, and hence we cannot compare them.



Year

Food Consumption per Capita, q

Food Price,

Consumer Income,

1927

88.9

91.7

57.7

1928

88.9

92.0

59.3

1929

89.1

93.1

62.0

1930

88.7

90.9

56.3

1931

88.0

82.3

52.7

1932

85.9

76.3

1933

86.0

78.3

43.8

1934

87.1

84.3

47.8

J935

85.4

88.1

52.1

1936

88.5

88.0

58.0

1937

88.4

88.4

59.8

1938

88.6

83.5

55.9

1939

91.7

82.4

60.3

1940

93.3

83.0

64.1

1941

95.1

86.2

73.7

(World War 11 years excluded)

1948

96.7

105.3

82.1

1949

96.7

102.0

83.1

1950

98.0

102.4

88.6

1951

96.1

105.4

88.3

1952

98.1

105.0

89.1

1953

99.1

102.6

92.1

1954

99.1

101.9

91.7

1955

99.8

100.8

1956

101.5

100.0

99.8

1957

99.9

99.8

99.9

1958

99.1

101.2

98.4

1959

101.0

98.8

101.8

1960

100.7

98.4

101.8

1961

100.8

98.8

103.1

1962

101.0

98.4

105.5

"1957-1959 = 100

Retail prices of Bureau of Labor Statistics, deflated by dividing by Consumer Price Index. Per capita disposable income, deflated by dividing by Consumer Price Index

The estimates of a, , and for 1927-1942 and 1948-1962 separately and for both periods combined are presented in Table 4.4. Note that we have presented the t-ratios in parentheses because this is more convenient to interpret. If we are interested in significance tests, the /-ratios are more convenient and if we are interested in obtaining confidence intervals, it is convenient to have the standard errors.

Table 4.3 Indexes of Food Consumption, Food Price, and Consumer Income"



[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [ 54 ] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]