TS-Summary and Themes1&2
TS-Summary and Themes1&2
(TIME SERIES)
CONTENT:
1. INTRODUCTION
3. ESTIMATION
4. HYPOTHESIS TESTING
5. MODEL SELECTION
6. SEASONALITY
7. PREDICTION
14. COINTEGRATION
Bibliography:
*. G.E.P. Box and G.M. Jenkins, Time series analysis. Forecasting and control. San
Francisco. Holden-Day (1970).
*. Aznar, A. y Trivez, F.J., Métodos de predicción en Economía II. Análisis de series temporales,
Ariel, Economía (1993).
1
2
1. INTRODUCTION
250
200
150
100
50
0
1947m1 2004m12
Time series análisis is not reduced to Economics. In fact, it is also studied in many other
areas such as Meteorology, Medicine, Biostatistics, Astronomy, Political Sciences, etc.
Find a mathematical model to explain the behaviour of the time series, in order to
Basic concepts:
3
Time series: y1, y2, y3, …, yT
We treat each observation (data) as if it were a random variable, such that we identify
the time series with the stochastic process
that is, each random variable (Yt), has behind a distribution function, such that
A common assumption in time series is to assume that all variables (that is, all the
observations are i.i.d., = independent and identically distributed).
Given that we treat all the observations as random variables, we need to define their
moments:
However, given that in most cases we only have a single realization available for each
of the random variables, we need to impose some restrictions to make these concepts
feasible. The basic concept here is the one called STATIONARITY.
* Stationarity
A stochastic process (or a time series) is said to be stationary (in a weak sense or
stationary of order 2) if it satisfies the following three properties:
4
In other words, the mean and the variance do not depend on time, and the covariance
between any two observations only depends on the distance between them, but not on
the specific location in time.
under stationarity,
Cov(y1, y4) = Cov(y2, y5) = Cov(y4, y7) = … = Cov(y10, y13) = etc. = γ3.
1 T
*. ̂ = y = yt
T t =1
1 T
*. ˆo = ( yt − y )
2
T t =1
1 T
*. ˆk = ( yt − y ) ( yt + k − y ).
T t =1
If the stochastic process is stationary and ergodic, these estimators, ̂ , ˆo , y ˆk will be
consistent estimators of μ, γo and γk. (That is, the estimators will tend to be more precise
as the sample size increases).
Though we do not properly define ergodicity, this property requires that observations
which are distant apart will tend to be uncorrelated. That is, a sufficient condition for
ergodicity is:
γk → 0 as k → ∞.
Note that:
F(y1, y2, y3, y4, …, yT) = F(y1+c, y2+c, y3+c, y4+c, …, yT+c) for all T and c,
5
where F represents the distribution function of probabilities.
Note that while the definition of weak stationarity is based on moments, the definition
of stationarity in strict sense is based on the distribution function.
Thus, the two definitions are independent and the only relation between them is given
by:
Across this course, we mainly work with stationary (in weak sense or order 2) series.
However, most of economic time series are not stationary, and, we will have to
transform them to render stationarity
5,4
5,2
4,8
4,6
4,4
4,2
4
1970m1 2004m12
1,55
1,5
1,45
1,4
1,35
1,3
1,25
1,2
1 T
6
2000
1000
-1000
-2000
7.Jan.86 31.Dec.97
The simplest example of a stationary process is the WHITE NOISE process. Generally,
it is denoted as εt.
2
4
1
2
0
0
-1
-2
-2
-3 -4
-4 -6
1 100 1 1000
If the time series is nonstationary, we will find a way to transform the time series in
another one that is stationary. Once the new series is found to be stationary, we will
7
look for the mathematical model that best describes the behaviour of the series,
such that the part that is un-modelled is precisely a white noise process..
Then,
yt = f(*) + εt,
where f is the mathematical function that explain the series and εt is a white noise
process.
and, in general,
Thus, we observe that the mean, the variance, and the covariance structure do not
depend on time, so the series yt is stationary.
If yt is stationary,
8
That is, the autocovariance function is symmetric, so we just focus on values of k ≥ 0.
Similarly, we define
k
k = = autocorrelation function
o
k = ( k ) = ( −k ) = − k .
In the previous example, (yt = εt + θ εt-1), we get the following values for the
autocorrelation structure (Exercise 1.3):
o = 1,
1 = ,
1 + 2
k = 0, for all k ≥ 2.
1 T
̂ = y = yt (sample mean)
T t =1
1 T
ˆo = 2
( yt − y ) (sample variance)
T t =1
1 T
ˆk = ( yt − y ) ( yt + k − y ). (sample covariance).
T t =1
ˆk
ˆ k = , k = 1, 2, ... (1.1)
ˆo
9
Thus, we have the sample autocorrelation function (Correlogram) → based on data
Example:
Suppose we have a theoretical model that is a white noise process (εt). Its
autocorrelation function (theoretical) will be given by the figure in the left-hand-side
below. However, let us suppose now that we have data, which have no structure, and we
compute its correlogram (sample autocorrelation function (1.1)). The values here (right-
hand-side plot), are similar to those given by the white noise model, and therefore we
can conclude that the time series follows a white noise process. (yt = εt)
1
1
0,8
0,8
0,6
0,6
0,4
0,4 0,2
0,2 0
-0,2
0 0 1 2 3 10
0 1 2 3 10
The correlogram tries to reproduce the theoretical autocorrelation function but it does
not imply to be the same. The correlogram will be the instrument employed to identify
the model.
Lk yt = yt-k
∆: First differences: ∆ = (1 – L)
∆ yt = (1 – L) yt = yt - yt-1
10
2. ARMA PROCESSES
The combination of the two processes leads to a new process called ARMA
(AutoRegressive + Moving Average).
Before starting with the processes, we define a fundamental Theorem in Time Series
Analysis:
“Any stationary process of order 2, yt, can be uniquely represented as the sum of two
processes mutually uncorrelated, (vt and wt), such that:
yt = vt + wt,
cs t − s ,
2
wt = with cs ,
s =0 s =0
In general (and in particular, in this chapter), we focus on the stochastic part (wt).
ARMA PROCESSES
yt = yt −1 + t , t = 1, 2, ... (2.1)
11
where εt is a white noise process.
yt = [ yt − 2 + t −1 ) + t = 2 yt − 2 + t + t −1, (2.3)
yt − 2 = yt −3 + t − 2 , t = 1, 2, ... (2.4)
yt = 2 [ yt − 3 + t − 2 ) + t + t −1 = 3 yt − 3 + t + t −1 + 2 t − 2
yt = n yt − n + t + t −1 + 2 t − 2 + ... + n −1 t − n +1,
or
n −1
yt = n yt − n + j t − j .
j =0
t − j ,
j
yt = (2.5)
j =0
since if 1, then n → 0 as n → ∞.
An alternative way of getting the same result is using the lag operator:
12
yt = yt −1 + t , t = 1, 2, ...
it can be represented as
yt = Lyt + t , t = 1, 2, ...
or
(1 − L) yt = t , t = 1, 2, ...
1
yt = t , t = 1, 2, ... (2.6)
(1 − L)
1
If │x│ < 1, 1 + x + x 2 + x 3 + ...
1− x
such that, if 1,
1
1 + L + 2 L2 + 3 L3 + ... (2.7)
1 − L
1
yt = t = (1 + L + 2 L2 + 3 L3 + ...) t
(1 − L)
yt = t + t −1 + 2 t − 2 + 3 t −3 + ...
t − j ,
j
yt =
j =0
1,
13
Assuming that the AR(1) process is stationary,
yt = yt −1 + t , 1,
E ( yt ) = E ( yt −1 + t ) = E ( yt −1 ) + E ( t ) .
Noting that in stationary processes, the mean is constant, (μ), and that E(εt) = 0, we get
= + 0 (1 − ) = 0 = E ( yt ) = 0.
again here, given that in stationary processes, the variance of yt is constant (and we
denote it by γo), looking at (2.8) we get:
o = 2 o + 2 + 0,
2
(1 − 2 ) o = 2 o = Var ( yt ) = .
(1 − 2 )
The last term in (2.9) is 0, given that a stationary AR(1) process depends exclusively on
εt, εt-1, εt-2, … and thus, it is uncorrelated with εt+1. Then, we obtain that:
1 = o = 2.
1− 2
2
2 = 1 = 2 o = 2.
1− 2
14
3
3 = 2 = 3 o = 2.
1− 2
k
k = k −1 = 2 k − 2 = ... = 2.
1− 2
o = 1;
1 o
1 = = = ;
o o
2 2 o
2 = = = 2;
o o
and, in general
k k o
k = = = k;
o o
and we see that the decay in the autocorrelation function follows an exponential process.
We first suppose that is a positive value. If that value is close to 1 ( = 0.9) the
decay in the autocorrelation function will be slow. On the contrary, if is close to 0,
(e.g. = 0.2), the decay is very fast, and the values become fast close to. On the other
hand, given that is positive, all the values in the autocorrelation function are positive.
If, on the contrary, is negative, we get the same exponential decay to 0, depending on
the magnitude of . However, such decayment oscillates, from postive to negative
values.
15
Funciones de autocorrelación para distintos procesos AR(1)
= 0.9 = 0.2
1 0,25
0,9
0,8 0,2
0,7
0,6 0,15
0,5
0,4 0,1
0,3
0,2 0,05
0,1
0 0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
= − 0.9 = − 0.2
1 0,1
0,8
0,05
0,6
0
0,4
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
0,2 -0,05
0 -0,1
-0,2 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
-0,15
-0,4
-0,2
-0,6
-0,8 -0,25
-1
Later on, we will study higher order autoregressive (AR) processes, e.g., AR(2), AR(3)
and in general AR(p) processes.
16
Given (2.10),
E ( yt ) = E ( t + t −1) = E ( t ) + E ( t −1) = 0.
2 + 2 2 = (1 + 2 ) 2 .
3 = 4 = 5 = ... = 0.
o = 1;
1 2
1 = = = ;
o (1 + 2 ) 2 (1 + 2 )
2 0
2 = = = 0;
o o
3 = 4 = 5 = ... = 0.
17
Finally, its graphical representation will consist on a unique value (positive or negative)
depending on the sing of θ.
= 0.9 = 0.2
0,6
0,6
0,5 0,5
0,4 0,4
0,3 0,3
0,2
0,2
0,1
0,1
0
0
1 3 5 7 9 11 13 15 17 19
1 3 5 7 9 11 13 15 17 19
= − 0.9 = − 0.2
0,6 0,6
0,5
0,4
0,4
0,2 0,3
0,2
0
0,1
1 3 5 7 9 11 13 15 17 19
-0,2 0
-0,1 1 3 5 7 9 11 13 15 17 19
-0,4
-0,2
-0,6 -0,3
Once more, if we are working with data, and the correlogram shows a plot similar to
one of these graphics, we can conclude that the time series follows a MA(1) process.
We have said before that the MA(1) processes (and in general any MA process) are
always stationary. Sometimes, it is desirable the process to be invertible, that means that
the process admits an AR(∞) representation. Then, it will require that < 1
(Invertibility condition). Then, it can be shown that,
yt = t + t −1
yt = (1 + L) t
1
yt = t
(1 + L)
18
1
yt = t (2.11)
(1 − (− ) L)
1
1 + (− ) L + (− ) 2 L2 + (− )3 L3 + ... = 1 − L + 2 L2 − 3 L3 + ....
1 − (− ) L
so (2.11) becomnes:
(1 − L + 2 L2 − 3 L3 + ....) yt = t
or
yt = yt −1 − 2 yt − 2 + 3 yt − 3 − .... + t ,
*. AR(2) processes
y t = 1 y t −1 + 2 y t −2 + t , t = 1, 2, ... (2.12)
y t = 1 L y t + 2 L2 y t + t , t = 1, 2, ...
(1 − 1 L − 2 L2 ) y t = t , t = 1, 2, ... (2.13)
1
yt = t , t = 1, 2, ...
(1 − 1 L − 2 L2 )
Next, we look at the conditions that are required for an AR(2) process to be stationary.
1 − 1 x − 2 x 2 = 0 .
Next, if we invert it (in the sense that we write which is in the right in the left and
viceversa) and substitute x by z, we get the following equation:
z 2 − 1 z − 2 = 0,
19
which is a second order equation (in standard form), and we can compute its roots, z1
and z2,
− b b2 − 4 a c 1 12 + 4 2
zi = = ,
2a 2
z1 1 y z2 1. (2.14)
In other words, the stationarity condition in AR(2) processes is that “the roots of the
“inverted” polynomial must be both smaller than 1 in absolute value”.
Example:
(1 − 0.2 L − L2 ) y t = t , t = 1, 2, ...
1 − 0.2 x − x 2 = 0.
z 2 − 0.2 z − 1 = 0,
Given that one of the roots (1.10) is above 1 in absolute value, the process given by
(2.15) is non-stationary.
----------------------------------------------------------------------------------------------------------
20
We have just said that the stationarity condition in AR(2) process is that the roots
− b b2 − 4 a c 1 12 + 4 2
zi = = ,
2a 2
z1 1 and z2 1.
z1 = a + b i
z2 = a - b i
zi = a2 + b 2 1.
Using this last relationship (2.16), we get that yt in (2.12) can also be expressed in the
following way:
1
yt = t ,
(1 − 1 L − 2 L2 )
21
1 1
y t = t ,
1 − z1 L 1 − z 2 L
y t = t + 1 t −1 + 2 t −2 + 3 t −3 + ....
where
1 = z1 + z 2 ,
2 = z12 + z 22 + z1 z 2 ,
Example:
z 2 + 0.2 z − 0.08 = 0,
22
− b b2 − 4 a c − 0.2 0.04 + 0.32 − 0.2 0.6
zi = = = = 0.2 and -
2a 2 2
0.4.
Given that the two roots are smaller than 1 in absolute value, the process is stationary.
1 1 1
yt = t = t =
(1 + 0.2 L − 0.08 L2 ) (1 − 0.2 L) (1 + 0.4 L)
such that
1. 1 + 2 1
2. 2 − 1 1
3. 2 − 1.
If one of these conditions is not satisfied, the process will not be stationary.
23
Thus, all the values of 1 and 2 that lie within the triangle will imply stationarity.
Moreover, if we draw, within the triangle, the curve 12 + 4 2 = 0 , we can also
distinguish the real roots from the complex ones.
yt = 1 yt −1 + 2 yt − 2 + t , t = 1, 2, ...
E ( yt ) = E (1 yt −1 + 2 yt − 2 + t ) = 1 E ( yt −1) + 2 E ( yt − 2 ) + E ( t )
24
= 1 + 2 (1 − 1 − 2 ) = 0 E ( yt ) = = 0.
2 + 2 1 2 1
Var ( y t ) = o = .
1 − 12 − 22
1
1 = 1 o + 2 1 (1 − 2 ) 1 = 1 o 1 = o.
1 − 2
2 = Cov( yt , yt − 2 ) = Cov(1 yt −1 + 2 yt − 2 + t , yt − 2 ) =
3 = 1 2 + 2 1,
25
and, in general,
s
= 1 s −1 + 2 s − 2 , s = 0, 1, ...
o o o
such that,
o = 1.
1
1 = .
1 − 2
12
2 = + 2 , etc.
1 − 2
These values can be obtained from (2.18), taking into account that ρk = ρ-k and ρo = 1.
1
0,8
0,8
0,4
0,6
0
0,4
-0,4
0,2
0 -0,8
0 10 20 0 10 20
26
*. MA(2) processes
yt = t + 1 t −1 + 2 t − 2 , t = 1, 2, ... (2.19)
or
yt = t + 1 L t + 2 L2 t , t = 1, 2, ...
yt = (1 + 1 L + 2 L2 ) t , t = 1, 2, ...
E ( yt ) = = 0.
Var ( yt ) = o = (1 + 12 + 22 ) 2 .
Cov ( yt , yt +1 ) = 1 = (1 + 1 2 ) 2 .
Cov( yt , yt + 2 ) = 2 = 2 2 .
o = 1.
1 + 1 2
1 = .
1 + 12 + 22
2
2 = .
1 + 12 + 22
Thus, the autocorrelation function of MA(2) processes is null for lags higher than 2.
Let’s see some examples in graphical terms:
27
Autocorrelation functions of MA(2) processes
1,2 1,2
1 0,8
0,8
0,4
0,6
0
0,4
-0,4
0,2
0 -0,8
0 1 2
*. ARMA(1, 1) processes
or
yt = L yt + t + L t , t = 1, 2, ...
(1 − L) yt = (1 + L) t , t = 1, 2, ...
stationary if: 1,
Given (2.20):
E ( yt ) = E ( yt −1 + t + t −1 ) = E ( yt −1 ) + E ( t ) + E ( t −1 ),
= + 0 + 0 (1 − ) = 0 E ( yt ) = = 0.
28
Var ( yt ) = Var ( yt −1 + t + t −1 ) = 2 Var ( yt −1 ) + Var ( t ) + 2 Var ( t −1 ) +
Again here, Cov( yt −1, t ) = 0, and Cov( t , t −1 ) = 0. We focus first on the last term,
Cov( y t −1 , t −1 ) = Cov( y t −2 + t −1 + t −2 , t −1 ) = 2 .
We get:
(1 + 2 + 2 ) 2
o = 2 o + 2 + 2 2 + 2 2 o = Var ( yt ) =
1 − 2
Cov( yt , t ).
1 = o + 2.
Cov( yt , t +1) = 1.
Similarly,
3 = Cov( yt , yt + 3 ) = 2 = 3 o + 2 2 ,
s = Cov( yt , yt + s ) = s −1 = s o + s −1 2 .
o = 1;
29
1 o + 2 2
1 = = = + ;
o o o
2 1
2 = = = 1;
o o
s
s = = s −1.
o
Thus, for s > 1, we get the same structure as in purely AR(1) processes. Generalizing all
the previous processes, we have
*. AR(p) processes
or
yt = 1 L yt + 2 L2 yt + ... + p L p yt + t , t = 1, 2, ...
(1 − 1 L − 2 L2 − ... − p L p ) yt = t , t = 1, 2, ...
p ( L) yt = t , t = 1, 2, ...
( z p − 1 z p −1 − 2 z p − 2 − ... − p )
z1 1; z2 1; .... zp 1.
Next we look at the autocovariance and the autocorrelation structures. Given (2.21):
30
E ( yt ) = 1 E ( yt −1 ) + 2 E ( yt − 2 ) + ... + p E ( yt − p ) + E ( t )
= 1 + 2 + ... + p + 0 (1 − 1 − 2 .... − p ) = 0 = 0.
Instead of looking at the variance, (γo), we look at the autocovariances for a given value
k. For this purpose, we first multiply both sides in (2.21) by yt-k, such that
yt yt − k = 1 yt −1 yt − k + 2 yt − 2 yt − k + ... + p yt − p yt − k + t yt − k ,
and next we calculate the mathematical expectation, (E(.)), in both sides of the previous
equality, that is,
E ( yt yt − k ) = 1 E ( yt −1 yt − k ) + 2 E ( yt − 2 yt − k ) + ... + p E ( yt − p yt − k ) + E ( t yt − k ).
Taking into account that the means are 0, we can re-write that expression as:
k = 1 k −1 + 2 k − 2 + ... + p k − p . (2.22)
k = 1 k −1 + 2 k − 2 + ... + p k − p ,
1 = 1 + 2 1 + ... + p p −1
2 = 1 1 + 2 + ... + p p − 2
3 = 1 2 + 2 1 + ... + p p − 3
...
p = 1 p −1 + 2 p − 2 + ... + p .
That is, we have p equations to solve p unknowns (ρ1, …, ρp), which is sufficient to
calculate the autocorrelation function. This system of equations is known as the YULE-
WALKER equations.
31
*. MA(q) processes
yt = t + 1 t −1 + 2 t − 2 + ... + q t − q , t = 1, 2, ...
or
yt = (1 + 1 L + 2 L2 + .... + q Lq ) t , t = 1, 2, ...
yt = q ( L) t , t = 1, 2, ....
( z q + 1 z q −1 + 2 z q − 2 + ... + q )
z1 1; z2 1; .... zq 1.
k =
2 (1 + 12 + 22 + ... + q2 ) if k = 0,
2 ( k + 1 k +1 + 2 k + 2 + ... + q k + q ) if k = 1, 2, …, q,
0 if k > q.
32
Thus, the autocorrelation function is given by:
k =
1 if k = 0,
( k + 1 k +1 + 2 k + 2 + ... + q k + q )
if k = 1, 2, …, q,
(1 + 12 + 22 + ... + q2 )
0 if k > q.
Thus, the autocorrelation function of a MA(q) process is null for lags above q. This is a
very useful results noting that when we compute the correlogram of a given time series,
if it is close to 0 from a given lag, we can easily identify it as a MA process. For
instance, if the correlogram adopts the graphic in the left (Series 1) we can suggest that
the model that is behind the data is a MA(5), while the correlogram in the right hand
side might correspond to a MA(2) process.
Series 1 Series 2
1,2 1,2
0,8
0,8
0,4
3
0 0,4
0 1 2 4 5 0 1 2
-0,4 4
0
-0,8 0 1 2
-1,2 -0,4
*. ARMA(p, q) process
or
33
yt = 1 L yt + .... + p L p yt + t + 1 L t + ... + q Lq t , t = 1, 2, ...
p ( L) yt = q ( L) t , t = 1, 2, ....
( z p − 1 z p −1 − 2 z p − 2 − ... − p )
are smaller than 1 in absolute value. Similarly, the process will be invertible if the q
roots of the polynomial:
( z q + 1 z q −1 + 2 z q − 2 + ... + q )
and k = Cov( yt yt − k ) =
such that k =
k = 1 k −1 + 2 k − 2 + ... + p k − p k = q + 1, q + 2, …
That is, for lags above q, the autocorrelation function does not take into account the
contribution from the MA part, and thus, its behaviour from lag q is the same as in an
AR(p) process.
34
*. Additional concepts in Chapter 2
We have seen before that it is quite easy to detect MA(q) processes from the
correlogram. In fact, it is simply to observe from which lag the values are close to 0.
Thus, for example, if the values in the correlogram are approximately 0 in lag 4 and in
all the following ones, the time series may be described as a MA(3) process.
However, to detect AR(p) processes from the correlogram is not an easy task since the
autocorrelation function substantially changes depending on the parameters of the
model Here, we can use another instrument that is called partial correlogram.
yt = 11 yt −1 + t ,
yt = 21 yt −1 + 22 yt − 2 + t ,
and sucesively up to
yt = p1 yt −1 + p 2 yt − 2 + ... + pp yt − p + t .
As you can see, all of them are AR processes with orders that are increasing from one
process to another. The sucession of values
is called the Partial Autocorrelation Function, and is formed by the series of the last AR
coefficients which are sucessively higher.
If, instead of ss , we use the estimators of such values, ˆss , the sucession of values
35
the time series will be an AR(p) process.
Example:
Let us suppose that the partial correlograms of the following series 3 and 4 adopt the
following form
:
Series 3 Series 4
1,2 1,2
0,8
0,8
0,4
3
0 0,4
0 1 2 4 5 0 1 2
-0,4 4
0
-0,8 0 1 2
-1,2 -0,4
Then, we can conclude by saying that Series 3 follows an AR(5) process, while Series 4
follows an AR(2) process.
Given the time series {y1, y2, y3, …, yT}, we can study its statistical properties from a
matritial viewpoint.
36
y1
y
y = 2
−
y
T
y N (m, )
(Tx1) (Tx1) (TxT )
where
E ( y1 ) 1
E ( y2 ) 2
m = =
− −
E( y )
T T
E ( y1 )
E ( y2 )
m =
− = − =
E( y )
T
37
V ( y1 ) Cov ( y1 y2 ) ... Cov ( y1 yT ) o 1 2 ... T −1
Cov ( y2 y1 ) V ( y2 ) ... Cov ( y2 yt ) 1 o 1 ... T − 2
= = ... .
... ... ... ... ... ... ... ...
Cov ( y y ) Cov ( y y )
T 1 T 2 ... V ( yT ) T −1 T − 2 ... ... o
This type of matrices, (where all the elements across the diagonals are equal) are called
TOEPLITZ matrices.
Adic. 3: Nonstationarity
At the beginning of this course we said that most economic time series are of a
nonstationary nature.
Here we briefly present two alterantive ways of transforming a time series that is
nonstationary into stationary.
A usual way (at least until the 80s) of describing the nonstationarity was to employ
deterministic functions (generally linear or quadratic) in time. That is, you can regress yt
on a variable “t” that describes a linear trend, or to regress it on the variables “t” y “t2”
(quadratic function), assuming that the resulting residuals are stationary. Thus, for
example, if the time series is of form
5,4
5,2
4,8
4,6
4,4
4,2
4
1970m1 2004m12
yt = + t + ut
such that once the parameters are estimated, ( ̂ y ̂ ) , the resulting residuals are
stationary and can be modelled using the ARMA models:
p ( L ) ut = q ( L ) t . (2.23)
38
In other words, what we do is to use a linear trend such that the discrepancies between
the observed values and the estimated ones form the linear trend are stationary. This
type of transformations is usually called the “deterministic approach”.
ut = (1 - L) yt = yt - yt-1 .
Assuming that ut is now stationary, we proceed to study its ARMA structure in the same
way as in (2.23). This approach is called the “stochastic approach” and will be studied
in detail in future chapters.
p ( L) (1 − L) d yt = q ( L) t .
where
p (L) is the AR polynomial of order p,
39
40