back start next


[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [ 53 ] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]


53

tl + d.f.

and d.f. = (n - 5) in this case (one a and four ps).

Partial correlations are very important in deciding whether or not to include more explanatory variables. For instance, suppose that we have two explanatory variables x, and x,, and rjj is very high, say 0.95, but , is very low, say 0.01. What this means is that if Xj alone is used to explain y, it can do a good job. But after x, is included, Xj does not help any more explaining y; that is, x, has done the job of X2. In this case there is no use including X2. In fact, we can have a situation where, for instance,

= 0.95 and /2 = 0.96

/,2 = 0.1 and , = 0.1

In this case each variable is highly correlated with but the partial correlations are both very low. This is called multicoUinearity and we will discuss this problem later in Chapter 7. In this example we can use x, only or Xj only or some combination of the two as an explanatory variable. For instance, suppose that X] is the amount of skilled labor, Xj the amount of unskilled labor, and the output. What the partial correlation coefficients suggest is that the separation of total labor into two components-skilled and unskilled-does not help us much in explaining output. So we might as well use x, + x or total labor as the explanatory variable.

4.6 Relationships Among Simple, Partial, and Multiple Correlation Coefficients

To study the relationships among the different types of correlation coefficients, we need to make use of the relationship that at each stage the residual sum of squares RSS = Syy{\ - R). Suppose that we have two explanatory variables X, and X2. Then

5vv(l - Rl 12) = residual sum of squares after both x, and X2 are included Syy{\ - ijt) = residual sum of squares after the inclusion of x, only

Now 1 measures the proportion of this residual sum of squares explained by X2. Hence the unexplained residual sum of squares after X2 is also included is

(1 - /•;2 ,)5 1 - rb)

in tiie multiple regression equation are all variables that are mentioned in the partial correlation coefficient. For instance, if we want to compute ; 237, we have to run a regression of on x, xi, x,, and x-,. Let 4 = P4/SE(P4), where P4 is the coefficient of x. Then

?4 237 =



Two Illustrative Examples

The following examples illustrate simple r, partial r, and R. The first example illustrates some problems in interpreting multiple regression coefficients when the explanatory variables are proportions. The second example illustrates the use of interaction terms.

Example 1: Hospital Costs

Consider the analysis of hospital costs by case mix by Feldstein. The data are for 177 hospitals. The explanatory variables are proportions of total patients treated in each category. There are nine categories: M = medical, P = pediatrics, S = general surgery, E = ENT, T = traumatic and orthopedic surgery, OS = other surgery, G = gynecology, Ob = Obstetrics. Other = miscellaneous others. The regression coefficients, their standard errors, r-values, partial rs, simple rs, and average cost per case are given in Table 4.2.

The entries in Table 4.2 need some explanation. The /-values are just the coefficients divided by the respective standard errors. The partial rs are obtained by the formula

/2 -I- d.f.

M. S.Feldstein, Economic Analysis for the Health Service Industry (Amsterdam: North-Holland, 1967), Chap. 1.

But this is equal to 5,1 - R] ,2). Hence we get the resuh that

1 - /?2,2 = (1 - f.Kl - 21)

If we have three explanatory variables we get

(1 - 2.,2 ) = (1 - rl,Kl - ,2.)(1 - .2) (4.12)

The subscripts, 1, 2, 3 can be interchanged in any required order. Thus if we need to consider them in order 3, 1, 2, we have

(1 - / ?,2 ) = (1 - ridO - /, )(1 - ,)

Note that the partial can be greater or smaller than the simpler r. For instance, the variable X2 might explain only 20% of the variance of y, but after Xi is included, it can explain 50% of the residual variance. In this case the simple ,-2 is 0.2 and the partial is 0.5. We are thus talking of the proportions explained of different variances.

On the other hand, the simple and the partial can never be greater than R (the square of the multiple correlation coefficient). This is clear from the formulas we have stated.



Table 4.2 Hospital Cost Regressions

Variable

Regression Coefficient

/ Ratio

Partial r

Simple r-

Average Cost per Case

44.97

18.89

2.38

0.0326

0.1423

114.48

-44.54

28.51

-1.56

0.0143

0.0074

24.97

-36.81

14.88

-2.47

0.0350

0.0343

32.70

-54.26

16.52

-3.28

0.0602

0.0947

15.25

-29.82

17.18

-1.74

0.0177

0.0062

39.69

28.51

20.27

1.41

0.0117

0.0478

98.02

- 10.79

21.47

-0.50

0.0015

0.0099

58.72

-34.63

16.34

-2.12

0.0261

0.0011

34.88

Other

69.51

Constant

69.51

= 0.3076



[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [ 53 ] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [155] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]