New Section 1
New Section 1
Save residuals means to extract difference of the true values of y and predicted values of
the regression.
Data Scaling
Functional form
For example because the coefficiant female -1.8 it means that women earn
1.8 units less than men
After logging the wage we understand the coefficient -0.3011 of female like this:
We can't put two dummy variables for woman and man due to perfect linear relationship they have
Multicollinearity assumption, then one of the variables would have to be ommited, it won't change
The result only the coefficient interpretation
In stata we could use the lincom command to fix the problem of the intercept
It might have different marginal effect since the difference in creditworthiness between
cr1 and cr2 might be different than the difference of credit worthiness between CR2
and CR3
Difference In slopes
The heteroscedasticity does not affect the estimators beta orr r^2
The F test is done on the U(auxiliary regression) but we don't cover it in this course and the statistic we
compare to the LM statistic(or BP statistic) is the chi squared critical value that we find using the
significance level and the independeng variables number.
Y creates biased estimates if everyone of lower income increses their salary in the survey
On the other hand the measurment error in the x is more severe because it lowers the explanitory ower
of x which couses the biased and inconsistant estimatesfor example ven we have error for the x it adds
the noise for the x which weakens the relationship between y and xthis leads to antuiation biaswhich
means that our coefficiants are biased
H is change in time
The mean and variance in
different time periods have to be
the same for the regression to be
stationary.
Weakly dependence- our current observations do not effect the future observations as the lag
h increases
The transformation is when we calculate the first difference since the regresion is non stationary or have
linear trends
White test
We use ar
F and t statistic????
Autoregressive conditional heteroskedasticity(when variance is not random but depend on past volatility)
Example
Dummy variables
Regression:
Since you suspect the relationship between x and y you create dummy variable
Nr and fr is from the example of the house values that are near incenirator and far from it
It looks how much the coefficient changet in between the pooled cross sectional models of different
times then we run the hypothesis H_0: coefficienat=0 and h1 h0 is not true using the t statistic
Panel data
This topic discusses two methods for estimating unobserved effects panel data
models besides first difference estimator.
Fixed effects estimator
While analyzing panel data fixed effects estimator Is another way to eliminate fixed
effets a_i by differencing the data across time periods
Fixed effects
Instumental variables estimation is another way how to fix the omitted variable problem. We
are using them when x is correlated with u term violating the zero conditional mean
assumption mlr4.
Mostly the same as in simple regresion only called two stages least squeres
In in sample we are looking at the r squared as an criteria the bigger r^2 the better it is
We use the data for out sample with data point that were not used in the in sample part.
Granger causality