Page 301 -
P. 301

300	 Chapter 7  Hypothesis Tests and Confidence Intervals in Multiple Regression

   ui given X1i and X2i—that is, vi = ui - E(ui ͉ X1i, X2i)—so that vi has conditional mean zero:
   E(vi ͉ X1i, X2i) = E3ui - E(ui ͉ X1i, X2i) ͉ X1i, X2i4 = E(ui ͉ X1i, X2i) - E(ui ͉ X1i, X2i) = 0 .
   Thus,

   Yi = b0 + b1X1i + b2X2i + ui

    = b0 + b1X1i + b2X2i + E(ui ͉ X1i, X2i) + vi (using the definition of vi)

	   = b0 + b1X1i + b2X2i + E(ui ͉ X2i) + vi (using conditional mean independence)
    = b0 + b1X1i + b2X2i + (g0 + g2X2i) + vi 3using linearity of E(ui ͉ X2i)4

    = (b0 + g0) + b1X1i + (b2 + g2)X2i + vi (collecting terms)

    = d0 + b1X1i + d2X2i + vi,                                                    (7.25)

		

   where d0 = b0 + g0 and d2 = b2 + g2.
         The error vi in Equation (7.25) has conditional mean zero; that is, E(vi͉ X1i, X2i) = 0.

   Therefore, the first least squares assumption for multiple regression applies to the final line

   of Equation (7.25), and if the other three least squares assumptions for multiple regression

   also hold, then the OLS regression of Yi on a constant, X1i, and X2i will yield unbiased and
   consistent estimators of d0, b1, and d2. Thus the OLS estimator of the coefficient on X1i is
   unbiased for the causal effect b1. However, the OLS estimator of the coefficient on X2i is
   not unbiased for b2 and instead estimates the sum of the causal effect b2 and the coefficient
   g2 arising from the correlation of the control variable X2i with the original error term ui.

         The derivation in Equation (7.25) works for any value of b2, including zero. A variable
   X2i is a useful control variable if conditional mean independence holds; it need not have a
   direct causal effect on Yi.

         The fourth line in Equation (7.25) uses the assumption that E(ui͉ X2i) is linear in X2i.
   As discussed in Section 2.4, this will be true if ui and X2i are jointly normally distributed.
   The assumption of linearity can be relaxed using methods discussed in Chapter 8. Exercise

   18.9 works through the steps in Equation (7.25) for multiple variables of interest and mul-

   tiple control variables.

         In terms of the example in Section 7.5 [the regression in Equation (7.19)], if X2i is
   LchPct, then b2 is the causal effect of the subsidized lunch program (b2 is positive if the
   program’s nutritional benefits improve test scores), g2 is negative because LchPct is nega-
   tively correlated with (controls for) omitted learning advantages that improve test scores,

   and d2 = b2 + g2 would be negative if the omitted variable bias contribution through g2
   outweights the positive causal effect b2.

         To better understand the conditional mean independence assumption, return to the con-

   cept of an ideal randomized controlled experiment. As discussed in Section 4.4, if X1i is ran-
   domly assigned, then in a regression of Yi on X1i, the conditional mean zero assumption holds.
   If, however, X1i is randomly assigned, conditional on another variable X2i, then the conditional
   mean independence assumption holds, but if X2i is correlated with ui, the conditional mean zero
   296   297   298   299   300   301   302   303   304   305   306