Ch6 Slides Ed3 Feb2024
Ch6 Slides Ed3 Feb2024
1
I. Time Series: Introduction
“Don’t never prophesy: if you prophesy right, nobody is going to remember and if
you prophesy wrong, nobody is going to let you forget”
Mark Twain
• So if the process is covariance stationary, all the variances are the same and all
the covariances depend on the difference between t1 and t2. The moments
E ( yt E ( yt ))( yt s E ( yt s )) s , s = 0,1,2, ...
are known as autocovariances.
• s approximately N(0,1/T) where T = sample size and s denotes the
autocorrelation at lag s
• We can use this to do significance tests for the autocorrelation coefficients
by constructing a confidence interval.
1
• For example, a 95% confidence interval would be given by .196
T.
• If the sample autocorrelation coefficient, s , falls outside this region for
any value of s, then we reject the null hypothesis that the true value of the
autocorrelation coefficient at lag s is zero.
Joint Hypothesis Tests: 1 =0, 2=0,…, m =0
(Data are independently distributed)
• We can also test the joint hypothesis that all m of the k autocorrelation
coefficients are simultaneously equal to zero using the Q-statistic developed
by Box and Pierce: m
Q T k2
k 1
where T = sample size, m = maximum lag length
• The Q-statistic is asymptotically distributed as a m.
2
• However, the Box Pierce test has poor small sample properties, so a variant
has been developed, called the Ljung-Box statistic:
m
k2
Q T T 2
~ m2
k 1 T k
• This statistic is very useful as a portmanteau (general) test of linear
dependence in time series.
An ACF Example
• Question:
Suppose that a researcher had estimated the first 5 autocorrelation coefficients
using a series of length 100 observations, and found them to be (from 1 to 5):
0.207, -0.013, 0.086, 0.005, -0.022.
a) Test each of the individual coefficient for significance
b) Use both the Box-Pierce and Ljung-Box tests to establish whether
they are jointly significant.
• Solution:
a) A coefficient would be significant if it lies outside (-0.196,+0.196) at the
5% level, so only the first autocorrelation coefficient is significant.
b) Q=5.09 and Q*=5.26
Compared with a tabulated 2(5)=11.07 at the 5% level, so the 5 coefficients
are jointly insignificant (low power of the joint test when 1 single
coefficient is significant).
III. Moving Average Processes
Var(Xt) = E[Xt-E(Xt)][Xt-E(Xt)]
but E(Xt) = 0, so
Var(Xt) = E[(Xt)(Xt)]
= E[(ut + 1ut-1+ 2ut-2)(ut + 1ut-1+ 2ut-2)]
= E[ u t2 12 u t21 22 u t2 2 +cross-products]
So Var(Xt) = 0= E [ ]
=
= (1 1 2 )
2 2 2
= 1 2 1 2 2
= ( 1 1 2 ) 2
Solution (cont’d)
2 = E[Xt-E(Xt)][Xt-2-E(Xt-2)]
= E[Xt][Xt-2]
= E[(ut +1ut-1+2ut-2)(ut-2 +1ut-3+2ut-4)]
= E[( 2 u t2 2 )]
= 2
2
3 = E[Xt-E(Xt)][Xt-3-E(Xt-3)]
= E[Xt][Xt-3]
= E[(ut +1ut-1+2ut-2)(ut-3 +1ut-4+2ut-5)]
=0
So s = 0 for s > 2.
Solution (cont’d)
(iii) For 1 = -0.5 and 2 = 0.25, substituting these into the formulae above
gives 1 = -0.476, 2 = 0.190.
IV. Autoregressive Processes
y t 1 y t 1 2 y t 2 ... p y t p u t
where ut is a white noise process with zero mean.
• Or using the lag operator notation:
Lyt = yt-1 Liyt = yt-i
p
y t i y t i u t
i 1
p
• or y t i Li y t u t
i 1
y t ( L)u t
where,
( L) (1 1 L 2 L2 ... p Lp ) 1
Characteristics of an Autoregressive Process
yt 1 yt 1 ut
(i) Given that the series is stationary, calculate the (unconditional)
mean of yt
• So kk measures the correlation between yt and yt-k after removing the effects
of yt-k+1 , yt-k+2 , …, yt-1 .
• The PACF is useful for telling the difference between an AR process and
an ARMA process.
• In the case of an AR(p), there are direct connections between yt and yt-s only
for s p.
where ( L) 1 1 L 2 L2 ... p Lp
or y t 1 y t 1 2 y t 2 ... p y t p 1u t 1 2 u t 2 ... q u t q u t
• Box and Jenkins (1970) were the first to approach the task of estimating
an ARMA model in a systematic manner. There are 3 steps to their
approach:
1. Identification
2. Estimation
3. Model diagnostic checking
Step 1: Identification
Step 2: Estimation
- Estimation of the parameters
- Can be done using least squares or maximum likelihood depending
on the model.
• The information criteria vary according to how stiff the penalty term is.
• The three most popular criteria are Akaike’s (1974) information criterion (AIC),
Schwarz’s (1978) Bayesian information criterion (SBIC), and the Hannan-
Quinn criterion (HQIC).
AIC ln( 2 ) 2 k / T
k
SBIC ln(ˆ 2 ) ln T
T
2k
HQIC ln(ˆ 2 ) ln(ln(T ))
T
where k = p + q + 1, T = sample size. So we min. IC s.t. p p, q q
SBIC embodies a stiffer penalty term than AIC.
• Which IC should be preferred if they suggest different model orders?
– SBIC is strongly consistent but inefficient.
– AIC is not consistent, but more efficient, and will typically pick “bigger”
models.
ARIMA Models