Data Analysis and Decision Making Time Series Analysis
Data Analysis and Decision Making Time Series Analysis
Making
Time Series Analysis
Time Series Analysis
Definition:-An ordered sequence of values of a variable at equally spaced time intervals .
Both of these goals require that the pattern of observed time series data is identified and
more or less formally described. Once the pattern is established, we can interpret and
integrate it with other data (i.e., use it in our theory of the investigated phenomenon, e.g.,
seasonal commodity prices). Regardless of the depth of our understanding and the validity
of our interpretation (theory) of the phenomenon, we can extrapolate the identified pattern
to predict future events.
Trend Analysis
There are no proven "automatic" techniques to identify trend components in the time series
data; however, as long as the trend is monotonous (consistently increasing or decreasing)
that part of data analysis is typically not very difficult. If the time series data contain
considerable error, then the first step in the process of trend identification is smoothing.
Smoothing:- Smoothing always involves some form of local averaging of data such that the
nonsystematic components of individual observations cancel each other out. The most
common technique is moving average smoothing which replaces each element of the series
by either the simple or weighted average of n surrounding elements, where n is the width of
the smoothing "window". Medians can be used instead of means. The main advantage of
median as compared to moving average smoothing is that its results are less biased by
outliers (within the smoothing window). Thus, if there are outliers in the data (e.g., due to
measurement errors), median smoothing typically produces smoother or at least more
"reliable" curves than moving average based on the same window width. The main
disadvantage of median smoothing is that in the absence of clear outliers it may produce
more "jagged" curves than moving average and it does not allow for weighting.
Page 2 of 5
In the relatively less common cases (in time series data), when the measurement error is
very large, the distance weighted least squares smoothing or negative exponentially
weighted smoothing techniques can be used. All those methods will filter out the noise and
convert the data into a smooth curve that is relatively unbiased by outliers.
Fitting a function:- Many monotonous time series data can be adequately approximated by
a linear function; if there is a clear monotonous nonlinear component, the data first need to
be transformed to remove the nonlinearity. Usually a logarithmic, exponential, or (less
often) polynomial function can be used.
Analysis of Seasonality
Seasonal dependency (seasonality) is another general component of the time series pattern.
It is formally defined as co relational dependency of order k between each i'th element of
the series and the (i-k)'th element and measured by auto correlation (i.e., a correlation
between the two terms); k is usually called the lag. If the measurement error is not too
large, seasonality can be visually identified in the series as a pattern that repeats
every k elements.
Page 3 of 5
Partial autocorrelations. Another useful method to examine serial dependencies is to
examine the partial autocorrelation function (PACF) - an extension of
autocorrelation, where the dependence on the intermediate elements (those within the lag)
is removed. In other words the partial autocorrelation is similar to autocorrelation, except
that when calculating it, the (auto) correlations with all the elements within the lag are
partialled out. If a lag of 1 is specified (i.e., there are no intermediate elements within the
lag), then the partial autocorrelation is equivalent to auto correlation. In a sense, the partial
autocorrelation provides a "cleaner" picture of serial dependencies for individual lags (not
confounded by other serial dependencies).
Page 4 of 5
Removing serial dependency:- Serial dependency for a particular lag of k can be removed by
differencing the series, that is converting each i'th element of the series into its difference
from the (i-k)''th element. There are two major reasons for such transformations.
First, we can identify the hidden nature of seasonal dependencies in the series. Remember
that, as mentioned in the previous paragraph, autocorrelations for consecutive lags are
interdependent. Therefore, removing some of the autocorrelations will change other auto
correlations, that is, it may eliminate them or it may make some other seasonalities more
apparent.
Page 5 of 5