back start next


[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [ 155 ] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]


155

plim P2 = P2 - , „2

These equations correspond to (11.4) considered earlier (which were for x2 = 0).

We will not present the detailed calculations here (because they are tedious), but the equations corresponding to (11.5) obtained from a regression of X on Y and Xl and normalizing with respect to Y are

plim p, = p, + I [P2MP2 + ppi - 22) + ol

(11.10)

I r.

where"

. A.*».pXl - 1- X,) + 2 2 = p,(l - - (p,X, - pp2x2)

[Substituting X2 « 0, we get the results in (11.5).]

If p, and p, are the estimators of p, and p, obtained from a regression of A2 on X, and Y, then plim p, and plim 02 are obtained by just interchanging the subscripts 1 and 2 in equations (11.10).

It is easy to see from equations (11.9) and (11.10) that the direction of biases are rather difficult to evaluate. Further, we saw eariier in the case of equations (11.4) and (11.5) that the derived bounds are in many cases rather too wide to be of any practical use and it is better to use equations (11.4) to get a range of consistent estimates based on different assumptions about X. The problem with equations (11.5) or (11.10) is that they depend on a further unknown parameter CT*.

where Thus

cov(„ w) = -P.Xi

2. W) = - 2 2

We thus have

a,, = cov(y, X,) = , + p2p -

- coVCF.y = p,p -bp, - P2X2 (11-8)

= ( ) = p? + Pi + 2pp,p2 + a - p?X, - p?X,

As before, we can derive the probability limits of p, and pj, the least squares estimators of p, and P2. We get

• Pl.mp.-P,-

p,X: - pp,X, I - p



In this case, we cannot get any bounds on the parameters using the procedures of calculating probability limits as in equations (11.9) and (11.10). There is an alternative procedure due to Klepper and Leamer" but its discussion is beyond our scope.

Here we will illustrate the range of consistent estimates of p, and p: using equations (11.8) or (11.9) with different assumptions about \, and Xj- This method is useful when we have some rough idea about the range of possible values for X, and X,. There is the problem, however, that not all these estimates are valid because the implied estimate of a- obtained from the third equation in (11.8) should be positive and also var(x,) var(jC2) - [cov(X, jc,)]- > 0. This last condition implies that (I - X,)(l - X2) - p- > 0. Using equations (11.9) we obtain consistent estimates pj and P2 of p, and p, by solving the equations

(1 - p - X,)p; + pX,32 = (1 - p2)p,

px.p; -1- (1 - p- - X2)p: = (1 - p)P2

The estimate of is obtained from the last equation in (11.8). It is

ff = - (pp + p; + 2pp;p; - x,p;- - X2P2-) (11.12)

As an example, let us consider the data in Table 7.3 and a regression of Y on X, and A",. The correlation coefficient between X, and X2 is 0.2156. Hence p = 0.2156. Also, a„ = 11.9378 and a,, = 3.2286.

It is reasonable to assume that A", (gross domestic production) has smaller measurement errors than A2 (stock formation). Hence we consider

X, = 0.01,0.02,0.05,0.10 •

X2 = 0.10.0.20, 0.30.0.40

Note that for all these values, the condition (I - X,)(I - X;) - p- > 0 is satisfied. The results of solving equations (II.11) are presented in Table I I.I. The range of consistent estimates of p, is (0.128, 0.201) and the range of consistent estimates of p, is (0.447. 0.697). Note that the OLS estimates were Pi = 0.191 and p, = 0.405. Thus with the assumptions about error variances we have made, the OLS estimate P2 is not within the consistent bounds for p, that we have obtained.

Note that assuming only A2 (and not A",) was measured with error we obtained earlier the bounds

0.169 < p, < 0.191 and 0.405 < P2 < 4.184

The OLS estimates are at the extremes of these bounds. With the assumptions about error variances we have made, the bounds for p, are wider but the bounds for p2 are much narrower.

The important conclusion that emerges from all this discussion and illustrative calculations is that in errors in variables models, making the most general

"S. Klepper and E. E. Leamer, "Consistent Sets of Estimates for Regressions with Errors in All Variables," Econometrica. Vol. 52, 1984, pp. 163-183.



0.01

0.10

0.1817

0.4519

155.5

0.01

0.20

0.1687

0.5119

155.5

0.01

0.30

0.1516

0.5903

155.5

0.01

0.40

0.1284

0.6970

155.4

0.02

0.10

0.1837

0.4514

155.5

0.02

0.20

0.1705

0.5114

155.5

0.02

0.30

0.1533

0.5898

155.5

0.02

0.40

0.1298

0.6965

155.4

0.05

0.10

0.1898

0.4500

155.5

0.05

0.20

0.1762

0.5099

155.5

0.05

0.30

0.1585

0.5882

155.5

0.05

0.40

0.1343

0.6949

155.4

0.10

0.10

0.2010

0.4473

155.5

0.10

0.20

0.1867

0.5071

155.5

0.10

0.30

0.1680

0.5853

155.5

0.10

0.40

0.1424

0.6920

155.4

assumptions about error variances leads to very wide bounds for the parameters, thus making no inference possible. On the other hand, making some plausible assumptions about the variances of the errors in the different variables, one could get more reasonable bounds for the parameters. A sensitivity analysis based on reasonable assumptions about error variances would be more helpful than obtaining bounds on very general assumptions.

11.4 Reverse Regression

In Sections 11.2 and 11.3 we considered two types of regressions. When we have the variables and x both measured with error (the observed values being Y and X), we consider two regression equations:

1. Regression of on X. which is called the "direct" regression.

2. Regression of X on Y. which is called the "reverse" regression.

Reverse regression has been frequently advocated in the case of analysis of salary discrimination." Since the problem is one of the usual errors in vari-

"There is a large amount of literature on this issue, many papers arguing in favor of "reverse regression." For a survey and difTerent alternative models, see A. S. Goldberger, "Reverse Regression and Salary Discrimination," Journal of Human Resources. Vol. 19. No. 3, 1984, pp. 293-318.

Table 11.1 Estimates of Regression Parameters and Error Variance Based on Different Assumptions About Error Variances



[start] [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35] [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62] [63] [64] [65] [66] [67] [68] [69] [70] [71] [72] [73] [74] [75] [76] [77] [78] [79] [80] [81] [82] [83] [84] [85] [86] [87] [88] [89] [90] [91] [92] [93] [94] [95] [96] [97] [98] [99] [100] [101] [102] [103] [104] [105] [106] [107] [108] [109] [110] [111] [112] [113] [114] [115] [116] [117] [118] [119] [120] [121] [122] [123] [124] [125] [126] [127] [128] [129] [130] [131] [132] [133] [134] [135] [136] [137] [138] [139] [140] [141] [142] [143] [144] [145] [146] [147] [148] [149] [150] [151] [152] [153] [154] [ 155 ] [156] [157] [158] [159] [160] [161] [162] [163] [164] [165] [166] [167] [168] [169] [170] [171] [172] [173] [174] [175] [176] [177] [178] [179] [180] [181] [182] [183] [184] [185] [186] [187] [188] [189] [190] [191] [192] [193] [194] [195] [196] [197] [198] [199] [200] [201] [202] [203] [204] [205] [206] [207] [208] [209] [210] [211] [212]