Eco Merged
Eco Merged
you can analyze the daily price of a product by observing its past four
weeks’ prices
AUTOCORRELATION
RAKESH SRIVASTAVA
DEPARTMENT OF STATISTICS
THE M.S.UNIVERSITY OF BARODA
VADODARA-390 002
PURPOSE:
such data are usually derived from the monthly data by simply adding three
monthly observations and dividing the sum by 3. this averaging introduces
smoothness into the data by dampening the fluctuations in the monthly data.
Therefore, the graph plotting the quarterly data looks much smoother than the
monthly data, and this smoothness may itself lend to systematic pattern in
disturbances , thereby introducing autocorrelation. Another source of
manipulation is interpolation or extrapolation of the data.
…(11)
…(12)
…(13)
CONT…
where cov(ut,ut+s) means covariance between error terms s periods apart and
where cor(ut,ut+s) means correlation between error terms s periods apart. Note
that symmetry property of covariances and correlations, cov(ut,ut+s)=cov(ut,ut-s)
and cor(ut,ut-s).
Since ρ is a constant between -1 and +1,in eq(11) shows that under scheme
AR(1) scheme, the variance of ut is still homoscedastic, but ut is correlated past. It
is critical to note that ׀ρ <׀1, that is, the absolute value of the rho is less than one.
If for example, rho is one, the variance and covariances listed above are not
defined. If ׀ρ <׀1, we say that the AR(1) process given in eq (9) is stationary;
that is mean, variance, covariance of ut do not change over time. If ׀ρ <׀1, then it
is clear from eq.(12) that the value of the covariance will decline as we go in the
past.
Now return to our two-variable regression model: Yt=β1+ β2Xt+ut . And we know
that the OLS estimators of slope coefficient is
…(14)
CONT…
() …(17)
The blue estimator in the presence of autocorrelation
In the two variable model, assuming that the AR(1) process, we can show
that the BLUE estimator of β2 is given by the following expression;
…(18)
…(19)
In this figure we show the 95% OLS [AR(1)] and GLS confidence interval
assuming that true = 0. consider a particular estimate of , say, b2. since b2
lies in the OLS confidence interval, we could accept the hypothesis that the true
is zero with 95% confidence but if we were to use the (correct) GLS confidence
interval, we could reject the null hypothesis that true is zero, for b2 lies in the
region of rejection.
The message is: to establish confidence intervals and to test hypothesis
one should use GLS and not OLS even though the estimators derived from
the latter are unbiased and consistent.
OLS ESTIMATION DISREGARDING AUTOCORRELATION
the situation is potentially very serious if we not only use but also continue
to use variance ( ), which completely disregards the problem of
autocorrelation, that is,, we mistakenly believe that the usual assumption of
classical linear model hold true. Errors will arise for the following reasons:
➢ The residual variance is likely to under estimate the
true .
➢ As a result, we are likely to over estimate R2.
➢ Even if is not under estimated, var ( ) may under estimate variance
, its variance under (first order) autocorrelation, even though the
latter is inefficient compare to var GLS.
➢ Therefore, the usual t and F tests of significance are no longer valid, and
if applied are likely to give seriously misleading conclusions about the
statistical significance of the estimated regression coefficients.
DETECTING AUTOCORRELATION
➢ Graphical Method
The assumption of non-autocorrelation of classical model relates to
population disturbances ut, which are not directly observable. What we have
instead are their proxies, the residuals , which can be obtained by usual OLS
procedure. Although the are not same thing as ut, very often usual
combination of ‘s give us some clue about likely presence of autocorrelation in
the u’s.
PATTERNS OF AUTOCORRELATION AND NONAUTOCORRELATION
CONT…
Figure (a) to (d) shows that there is discernible pattern among the u’s. Figure (a)
shows a cyclical pattern; figure (b) and (c) suggests an upward or downward
trend in disturbances; whereas figure (d) indicates that both quadratic and
linear trend in the disturbances. Only figure (e) indicates no systematic pattern,
supporting the nonautocorrelation assumption of classical linear regression
model.
Residuals and standardized residuals
There are various ways of examining the residuals. We can simply plot them
against time, the time sequence plot, as we have done in above Figure.
Alternatively, we can plot the standardized residuals against time, which are
also shown in above Figure. The standardized residuals are simply the
residuals divided by the standard error of the regression
that is, they are Notice that are measured in the
units in which the regress and Y is measured. The values of the standardized
residuals will therefore be pure numbers (devoid of units of measurement)
and can be compared with the standardized residuals of other regressions.
Moreover, the standatrized residuals, like , have zero mean and
approximately unit variance. In large samples
is approximately normally distributed with zero mean and unit variance.
Now we will plot against that is, plot the residuals at time “t” against
their value at time (t-1) [ in below figure], a kind of empirical test of the AR(1)
scheme. The figure reveals, that most of the residuals are bunched in the second
(northeast) and the fourth (southeast) quadrants, suggesting a strong positive
correlation in the residuals.
CONT…
➢ Durbin-Watson dTest
The most celebrated test for detecting serial correlation is that developed by
statisticians Durbin and Watson. It is popularly known as Durbin-Watson d
statistic, which is defined as
Where
That is asymptotically, n-p times the R2 value obtain from the auxiliary
regression (1.4) follows the chi-square distribution with p df. If in application
(n-p)R2 exceeds the critical chi-square value at the chosen level of significance,
we reject the null hypothesis, in which case at least one of the rho in (1.2) is
statistically significantly different from zero.
The following practical points about the BG TEST may be noted:
➢ The regressors included in the regression model may contain lagged values
of the regressand Y, that is, Yt-1, Yt-2, etc., may appear as explanatory
variables. Contrast this model with the DWD test restriction there be no
lagged value of the regressand among the regressors.
➢ As noted earlier, the BG test is applicable even if the distubances follow a pth
order moving average (MA) process, that is, the ut are generated as follows
ut = εt + λ1ε t-1+ λ2 ε t-2+…+ λp ε t-p
where εt is a white noise error term, that is, the error term that satisfies all the
classical assumptions.
CONT…
➢ If in the equation (1.2) p=1, meaning first order autocorrelation, then the BG
test is known as Durbin’s M test.
➢ A drawback of the BG test is that the value of p, the length no of the lag,
cannot be specified a prior. Some experimentation with the p value is
inevitable. Sometime one can use the so called Akaike and Schwarz
information criterion to select lag length.
REMEDIAL MEASURES
If after applying one or more of the diagnostic tests of autocorrelation discussed
in the previous section, we find that there is autocorrelation, what then? We have
four option:
1. Try to find out if autocorrelation is pure autocorrelation, not the result of
mis-specification of the model.
2. If it is pure autocorrelation, one can use appropriate transformation of the
original model so that in the transformed model we do not have the problem of
(pure) autocorrelation.
3. In the large samples, we can use the Newey-West method to obtain
standard errors of OLS estimators that are corrected for autocorrelation. This
method is actually an extension of White’s hetroscedasticity-consistent standard
errors method.
4. In some situations we can continue to use the OLS method.
MODEL MIS-SPECIFICATION VERSUS PURE
AUTOCORRELATION
1) ρ is known
If the coefficient of first order autocorrelation is known, the problem of
autocorrelation can be easily solved if equation (2.1) holds true at time t it is
also holds true at time (t-1). Hence
…(2.2 )
multiplying equation (2.2) by ρ on both sides we obtain
…(2.3)
subtracting eq. (2.3) from (2.1)
…(2.4)
where εt = ut - ρut-1
We can express eq. (2.4) as
…(2.5)
Where
Regression (2.4) is known as Generalised or Quasi, Differenced equation. It
involves regressing Y on X, not in original form but in difference form which
is obtain by substituting a proportion (=ρ) of the value of variable in the
previous time period from its value in the current time period.
2) ρ is not known:-
The Cochrane-Orcutt iterative procedure to estimate ρ.:-
An alternative to estimating ρ from the Durbin-Watson d is the
frequently used Cochrane-Orcutt method that uses the estimated
residuals to obtain information about the unknown ρ.
To explain the method, consider the two-variable model:
…..(3.1)
…..(3.2)
Cochrane and Orcutt then recommend the following steps to
estimate ρ:
1) Estimate the two-variable model by the standard OLS routine and obtain
the residuals, .
2) Using the estimated residuals, run the following regression:
…..(3.3)
which is empirical counterpart of the AR(1) scheme given previously.
3) Using obtained from eq. 1, run the generalized difference equation,
namely,
Or,
……(3.4)
4) Since a priori it is not known that the obtained from eq.(3.3) is the
“best” estimate of ρ, substitute the values of and obtained
from eq. (3.4) into the original regression [i.e., in eq.(3.1)] and obtain the
residuals, say , as
….(3.5)
Which can be easily computed since are all known.
5) Now to estimate the regression
….(3.6)
Which is similar to eq.(3.3). thus is the second- round estimate of ρ.
Since we do not know whether this second round estimate of ρ is the best of
ρ, we can go into the third round estimate, and so on. As the preceding steps
suggests, the Cochrane-Orcutt method is iterative. But how long should we go
on? The general procedure is to stop carry out the
iterations when the successive estimates of ρ differ by very small amount,
say, by less than 0.01 or 0.005.
The Cochrane-Orcutt Two-Step Procedure:-
This is shortened version of the iterative process. In step one we estimate ρ from
the first iteration, that is, that is from regression (eq. 3.3), and in step two we use
that estimate of ρ to run the generalized difference equation. Sometimes in
practice this two step method gives results quite similar to those obtained from
the more elaborate iterative procedure discussed above.
Durbin’s two step method of estimating ρ:-
To illustrate this method, let us write the generalized difference equation
equivalently as
…..(3.7)
Durbin suggests the following two-step procedure to estimating ρ:
1) Treat eq. 3.7 as a multiple regression model, regressing Yt on Xt,
Xt-1, and Yt-1, and treat the estimated value of the regression coefficient of Yt-
1(= as an estimate of ρ. Although biased, it provides consistent estimate
of ρ.
2) Having obtained , transform the variables as
and and run the OLS regression on the transformed
variables as in
Where
MODEL
MISSPECIFICATION
➢ One of the assumption of the classical linear regression
model (CLRM), in that the regression model used in
analysis is “correctly” specified: if the model is not
correctly specified, we encounter the problem of model
specification error or model specification bias.
MODEL SELECTION CRITERIA
➢ Data admissible.
It should reflect data generating process and Prediction made from that model
must be logically possible.
➢ Consistent with theory.
It must make good economic sense.
➢ Have weakly exogenous regression.
The explanatory variables, or regressors must be uncorrelated with the error term.
➢ Exhibit parameter constancy.
The value of parameter should me stable.
➢ Exhibit data coherency.
The residuals estimated from the model must be purely random.
➢ Encompassing.
The model should encompass or include all the rival models in the sense that it is
capable of explaining their results.
➢ Parsimonious.
Model should be compact.
TYPES OF SPECIFICATTION ERROR
Assume that on the basis of the criteria just listed, we arrive at a
model that we accept a good model. To be concrete, let the model to
be
3.1
where Y = total cost of production and X = output. Equation 3.1 is
the familiar textbook example of the cubic total cost function.
But suppose for some reason, a researcher decides to use
following model:
if 3.1 is the true, 3.4 also constitutes a specification error, the error
here consisting in including an unnecessary or irrelevant variable
in the sense that true assumes to be zero. The now error term is in
fact
where the error term enters additively. Although the variables are the
same in the two models, we have denoted the slope coefficient in 3.8
by β and the slope coefficient in 3.9 by α. Now if 3.8 is the correct or
true model, would the estimated α provide an unbiased estimate of the
true β? This is, will If that is not the case, improper
stochastic specification of the error term will constitute another
source of specification error.
To sum up, in developing an empirical model, one likely to
commit one or more of the following specification errors:
CONSEQUENCES OF MODEL SSPECIFICATION ERROR
3.14
Model (3.13)
using 3.18
Thus, the explanatory variable and the error term in 3.20 are correlated,
which violates the crucial assumption of the classical linear regression
model that the explanatory variable in uncorrelated with the stochastic
disturbance term. If this assumption is violated, it can be shown that the
OLS estimators are not only biased but also inconsistent, that is, they are
remain biased even if the sample size n increases indefinitely.