M05 StockWatson123635 03 Econ Ch05
M05 StockWatson123635 03 Econ Ch05
Regression with a
Single Regressor:
Hypothesis Tests
and Confidence
Intervals
v2
~ N 1 , 2 2
, where 𝑣𝑖 = (𝑋𝑖 – 𝜇𝑋)𝑢𝑖
n( X )
General setup
Null hypothesis and two-sided alternative:
H0: β1 = β1,0 vs. H1: β1 ≠ β1,0
where β1,0 is the hypothesized value under the null.
var[( X i x )ui ] v2
var( )= = , where vi = (Xi – μX)ui.
n( 2X )2 n( X )
2 2
n (Xi X )
2
= (–3.30, –1.26)
so:
TestScore = 698.9 – 2.28×STR, , R2 = .05, SER = 18.6
(10.4) (0.52)
t (β1 = 0) = –4.38, p-value = 0.000 (2-sided)
95% 2-sided conf. interval for β1 is (–3.30, –1.26)
Copyright © 2011 Pearson Addison-Wesley. All rights reserved.
5-16
Summary of statistical inference about β0 and β1
Estimation:
• OLS estimators and
• and have approximately normal sampling distributions in
large samples
Testing:
• H0: β1 = β1,0 v. β1 ≠ β1,0 (β1,0 is the value of β1 under H0)
• t = ( – β1,0)/SE( )
• p-value = area under standard normal outside tact (large n)
Confidence Intervals:
• 95% confidence interval for β1 is { ± 1.96×SE( )}
• This is the set of β1 that is not rejected at the 5% level
• The 95% CI contains the true β1 in 95% of all samples.
When Xi = 0, Yi = β0 + ui
• the mean of Yi is β0
• that is, E(Yi|Xi=0) = β0
When Xi = 1, Yi = β0 + β1 + ui
• the mean of Yi is β0 + β1
• that is, E(Yi|Xi=1) = β0 + β1
so:
β1 = E(Yi|Xi=1) – E(Yi|Xi=0)
= population difference in group means
0 if STRi 20
OLS regression: TestScore = 650.0 + 7.4×D
(1.3) (1.8)
Tabulation of group means:
Yi = β0 + β1Xi + ui
• β0 = mean of Y when X = 0
• β0 + β1 = mean of Y when X = 1
• β1 = difference in group means, X =1 minus X = 0
• SE( ) has the usual interpretation
• t-statistics, confidence intervals constructed as usual
• This is another way (an easy way) to do difference-in-means
analysis
• The regression formulation is especially useful when we have
additional regressors (as we will very soon)
1. What…?
2. Consequences of homoskedasticity
3. Implication for computing standard errors
Heteroskedastic or homoskedastic?
Copyright © 2011 Pearson Addison-Wesley. All rights reserved.
5-26
The class size data:
Heteroskedastic or homoskedastic?
Copyright © 2011 Pearson Addison-Wesley. All rights reserved.
5-27
So far we have (without saying so)
assumed that u might be heteroskedastic.
• You can prove that OLS has the lowest variance among
estimators that are linear in Y… a result called the Gauss-
Markov theorem that we will return to shortly.
• The formula for the variance of and the OLS standard
error simplifies: If var(ui|Xi=x) = , then
var[( X i x )ui ]
var( )= (general formula)
n( ) 2 2
X
u2
= (simplification if u is homoscedastic)
n X 2
Comments
• The GM theorem is proven in SW Appendix 5.2
( X X )ui
1 n
= wi ui ,
i
– β1 = i1
n
n i1
( X i
X ) 2
i1
( Xi X )
where wi = n .
1
n i1
( X i X )2