Chapter 6 Time series
Learning Objectives
- Explain the concept and general properties of stationary, \(I(0)\), and integrated, \(I(1)\), univariate time series.
- Explain the concept of a stationary random series.
- Explain the concept of a filter applied to a stationary random series.
- Know the notation for backwards shift operator, backwards difference operator, and the concept of roots of the characteristic equation of time series.
- Explain the concepts and basic properties of autoregressive (AR), moving average (MA), autoregressive moving average (ARMA) and autoregressive integrated moving average (ARIMA) time series.
- Explain the concept and properties of discrete random walks and random walks with normally distributed increments, both with and without drift.
- Explain the basic concept of a multivariate autoregressive model.
- Explain the concept of cointegrated time series.
6.1 Theory
6.1.1 Time Series: A Deep Dive for CS2 Actuarial Professionals
Time series analysis is a cornerstone of risk modelling and survival analysis, enabling actuaries to understand and forecast phenomena that evolve over time. At its heart, a time series is a sequence of observations indexed in time order. Unlike independent observations, the temporal ordering of data in a time series is paramount, as the observations are intrinsically related to one another. This field of study models these sequences as realisations of stochastic processes, specifically those indexed in discrete time with a continuous state space. The primary goals of time series analysis include describing the data, constructing appropriate models, forecasting future values, and identifying when a process is out of control.
Let’s delve into the specific learning objectives:
6.1.1.1 1. & 2. Explaining the Concept and General Properties of Stationary, I(0), and Integrated, I(1), Univariate Time Series, and the Concept of a Stationary Random Series.
The concept of stationarity is fundamental to time series analysis, primarily because efficient model calibration can typically only be performed on stationary processes.
Strictly Stationary Process: A stochastic process
X
is considered strictly stationary if the joint probability distributions of any set of observations,X_t1, X_t2, ..., X_tn
, are identical to those ofX_{t1+k}, X_{t2+k}, ..., X_{tn+k}
for allt
in the time setJ
and any integerk
. This implies that all statistical properties of the process, such as probabilities, expected values, and variances, remain unchanged as time elapses.Weakly Stationary Process: A stochastic process is weakly stationary if it satisfies two less stringent conditions:
- The mean of the process,
E[X_t]
, is constant and does not depend on timet
. - The covariance between any two observations,
cov(X_s, X_t)
, depends only on the time difference or lag,t-s
(ork
ift-s = k
), and not on the absolute time pointss
ort
. As variance is a special case of covariance (var(X_t) = cov(X_t, X_t)
), it also must be constant for a weakly stationary process.
- The mean of the process,
Relationship between Strict and Weak Stationarity: If a process is strictly stationary, it is automatically weakly stationary. However, the reverse is not necessarily true. An important exception is multivariate normal processes, where strict and weak stationarity are equivalent because their distribution is fully determined by their mean vector and covariance matrix. In the study of time series, “stationary” often serves as a shorthand for “weakly stationary”.
Purely Indeterministic Processes: For a time series process to be truly considered “stationary” in our context, it must also be “purely indeterministic”. This means that the predictive power of past observations,
X_1, ..., X_n
, for a future valueX_N
diminishes asN
approaches infinity. This excludes deterministic series (e.g., repeating patterns) from our primary focus.Integrated Processes – I(d) Notation:
- I(0) Process: A time series process
X
is denoted asI(0)
if it is, by itself, a stationary time series process. - I(1) Process: A process
X
is denoted asI(1)
if it is non-stationary, but its first difference,∇X = X_t - X_{t-1}
, is a stationary process. - I(d) Process: More generally, a process
X
is integrated of orderd
, orI(d)
, if itsd
-th difference,∇^d X
, is a stationary process. Non-stationary random processes must be transformed into stationary ones (often by differencing) before model calibration can be efficiently performed.
- I(0) Process: A time series process
6.1.1.2 3. Explaining the Concept of a Filter Applied to a Stationary Random Series.
A filter, represented by a collection of weights a_k
, is applied to an input series x_t
to produce a modified output series y_t
, defined as y_t = Σ_{k=-∞}^{∞} a_k x_{t-k}
. The purpose of applying a filter is to modify the input series to achieve specific objectives or to highlight certain features of the data. For instance, filters are crucial in economic time series analysis for detecting, isolating, and removing deterministic trends. In practice, such filters usually consist of a relatively small number of non-zero components.
6.1.1.3 4. Knowing the Notation for Backwards Shift Operator, Backwards Difference Operator, and the Concept of Roots of the Characteristic Equation of Time Series.
These operators and the concept of characteristic roots are essential tools for manipulating and understanding time series models.
Backward Shift Operator (B): This operator acts on a time series
X_t
to give its value at the previous time point:B X_t = X_{t-1}
. It can be applied multiple times, for example,B^2 X_t = X_{t-2}
, and is useful for compacting time series equations. In R, differenced values usingB
can be generated withdiff(<time series>, lag=1, differences=1)
.Backward Difference Operator (∇): Defined as
∇ = 1 - B
, this operator calculates the difference between the current value and the previous value of a time series:∇ X_t = X_t - X_{t-1}
. This operator is particularly important for transforming non-stationary series into stationary ones. Repeated applications are also possible, such as∇^2 X_t = X_t - 2X_{t-1} + X_{t-2}
. For seasonal differencing, such as monthly data with a period of 12, the notation∇_12
meansX_t - X_{t-12}
, and the R command would bediff(<time series>, lag=12, differences=1)
.Roots of the Characteristic Equation: This concept is vital for determining the stationarity of Autoregressive (AR) models and the invertibility of Moving Average (MA) models.
- For Stationarity (AR Part): To test if an AR process is stationary, you construct a characteristic polynomial from its autoregressive terms. This is done by replacing
X_t
with1
,X_{t-1}
withλ
,X_{t-2}
withλ^2
, and so on. Setting this polynomial equal to zero, you then find its roots. The time series is stationary if all the roots are strictly greater than 1 in magnitude. - For Invertibility (MA Part): To test if an MA process is invertible, you follow a similar procedure, but using the white noise terms. Replace
e_t
with1
,e_{t-1}
withλ
,e_{t-2}
withλ^2
, and so on. Set this polynomial to zero and find its roots. The time series is invertible if all the roots are strictly greater than 1 in magnitude.
- For Stationarity (AR Part): To test if an AR process is stationary, you construct a characteristic polynomial from its autoregressive terms. This is done by replacing
6.1.1.4 5. Explaining the Concepts and Basic Properties of Autoregressive (AR), Moving Average (MA), Autoregressive Moving Average (ARMA) and Autoregressive Integrated Moving Average (ARIMA) Time Series.
These models form the backbone of linear time series analysis, particularly within the Box-Jenkins framework.
White Noise Process: Before discussing AR and MA models, it’s crucial to understand white noise. A white noise process is a sequence of independent and identically distributed (IID) random variables, typically assumed to have a mean of zero and a constant variance (
σ^2
). White noise processes are inherently weakly stationary and possess the Markov property in a trivial sense, as the future development is entirely independent of the past.Autocovariance Function (ACF - γ_k): For a stationary time series
X
, the autocovariance function measures the covariance betweenX_t
andX_{t-k}
(at lagk
):γ_k = cov(X_t, X_{t-k})
. Note thatγ_0
is simply the variance ofX_t
.Autocorrelation Function (ACF - ρ_k): The autocorrelation function is the normalized autocovariance, given by
ρ_k = γ_k / γ_0
. For purely indeterministic processes,ρ_k
tends to zero as the lagk
approaches infinity, indicating a diminishing connection between terms further apart.Partial Autocorrelation Function (PACF - φ_k): The PACF measures the conditional correlation between
X_{t+k}
andX_t
, given all the intermediate observationsX_{t+1}, ..., X_{t+k-1}
. Formulae forφ_k
are available in the Tables. For stationary ARMA processes,φ_k
will decay towards zero ask
approaches infinity.Autoregressive (AR) Process: An
AR(p)
process models the current value of a time series,X_t
, as a linear combination of itsp
past values and a white noise error terme_t
. The general form isX_t = μ + α_1(X_{t-1}-μ) + ... + α_p(X_{t-p}-μ) + e_t
.- Properties:
AR(p)
processes are stationary if the roots of their characteristic equation (derived from the AR part) are strictly greater than 1 in magnitude. They are always invertible. OnlyAR(1)
processes possess the Markov property. The ACF of anAR(p)
process decays geometrically, while its PACF cuts off (becomes zero) after lagp
. R’sarima.sim()
can simulate AR models, andarima()
can fit them.
- Properties:
Moving Average (MA) Process: An
MA(q)
process modelsX_t
as a linear combination of the current andq
past white noise error terms. The general form isX_t = μ + e_t + β_1 e_{t-1} + ... + β_q e_{t-q}
. This is often described as “smoothed noise”.- Properties:
MA(q)
processes are always stationary because they are a finite linear combination of stationary white noise terms. They are invertible if the roots of their characteristic equation (derived from the MA part) are strictly greater than 1 in magnitude.MA
processes are never Markov. The ACF of anMA(q)
process cuts off after lagq
, while its PACF decays geometrically.
- Properties:
Autoregressive Moving Average (ARMA) Process: An
ARMA(p,q)
process combines both autoregressive and moving average components. Its defining equation incorporates both pastX
values and paste
values.- Properties:
ARMA(p,q)
processes are stationary if the roots of the AR part’s characteristic equation are strictly greater than 1 in magnitude. They are invertible if the roots of the MA part’s characteristic equation are strictly greater than 1 in magnitude. The onlyARMA
process that is Markov isARMA(1,0)
, which is equivalent toAR(1)
. Both the ACF and PACF of a stationaryARMA(p,q)
process decay exponentially as the lag increases. Thearima()
function in R can be used to fit ARMA models.
- Properties:
Autoregressive Integrated Moving Average (ARIMA) Process: An
ARIMA(p,d,q)
process is a powerful generalization where thed
-th difference of the time series,∇^d X_t
, is a stationaryARMA(p,q)
process. This model is central to the Box-Jenkins Methodology for time series analysis.- Box-Jenkins Methodology: This systematic approach involves four main stages:
- Identification: Determine suitable orders (
p, d, q
) for theARIMA
model. This involves plotting the time series to detect trends or seasonal cycles and inspecting the Sample ACF (SACF) and Sample PACF (SPACF) plots. Thed
parameter (differencing order) is chosen to make the series stationary. Statistical tests like the Phillips-Perron (PP) test can confirm stationarity by checking for unit roots (PP.test(Xt)
in R). - Estimation: Estimate the model parameters (e.g.,
α
for AR,β
for MA) using methods like Maximum Likelihood Estimation (MLE) or the Method of Moments (Yule-Walker equations). - Diagnosis: Check the goodness of fit by analysing the residuals (the differences between observed and fitted values). The residuals should ideally resemble white noise. Diagnostic tests include visual inspection of residual plots, their SACF/SPACF, and formal statistical tests like the Ljung-Box (or “portmanteau”) test (
Box.test()
in R). The Akaike Information Criterion (AIC) is also used for model selection (AIC()
forarima()
objects). The turning points test, though not explicitly in Paper B Core Reading, is another diagnostic tool for residuals. - Forecasting: Once a satisfactory model is identified and estimated, it can be used to predict future values. For
ARIMA
processes withd > 0
, the prediction variance will increase to infinity as the forecast horizon lengthens. R’spredict()
function is commonly used for forecasting fromarima
objects.
- Identification: Determine suitable orders (
- Box-Jenkins Methodology: This systematic approach involves four main stages:
6.1.1.5 6. Explaining the Concept and Properties of Discrete Random Walks and Random Walks with Normally Distributed Increments, Both with and Without Drift.
Random walks are a fundamental class of non-stationary stochastic processes.
- General Random Walk: A general random walk,
X_n
, is defined asX_n = X_{n-1} + Y_n
(orX_n = Y_1 + Y_2 + ... + Y_n
withX_0 = 0
), whereY_n
are independent and identically distributed (IID) random variables. The time set is discrete. The state space can be discrete or continuous, depending on the nature of theY_n
(the “steps”).- Properties: A general random walk is not stationary because its mean and variance typically increase linearly with time. For example, if
E[Y_n] = μ_Y
(a non-zero drift), thenE[X_n] = nμ_Y
, which is not constant. Ifvar(Y_n) = σ_Y^2
, thenvar(X_n) = nσ_Y^2
, also not constant. - Markov Property: Despite being non-stationary, random walks generally possess the Markov property, as the next step (and thus the future state) depends only on the current position, not on how that position was reached.
- Integrated Process: A random walk
X_t = X_{t-1} + e_t
(wheree_t
is white noise) is anARIMA(0,1,0)
process because its first difference,∇X_t = e_t
, is stationary white noise.
- Properties: A general random walk is not stationary because its mean and variance typically increase linearly with time. For example, if
- Simple Random Walk: This is a special case of a general random walk where each
Y_j
(step) can only take values of +1 or -1.- Simple Symmetric Random Walk: A further specialization where the probability of taking a +1 step equals the probability of taking a -1 step, typically 0.5 for each.
- State Space and Time Set: For a simple random walk, the state space is typically the discrete set of integers
Z = {..., -2, -1, 0, 1, 2, ...}
, and the time set is the discrete set of non-negative integersJ = {0, 1, 2, ...}
. - Applications: Examples include tracking a player’s profit/loss in a simple casino game or modelling No Claims Discount (NCD) levels in motor insurance, although the latter often involves more complex transition rules.
- Boundaries: Random walks can be further characterized by their boundaries:
- Absorbing Barrier: A state from which the process cannot leave. If a random walk enters an absorbing state, it remains there forever.
- Reflecting Boundary: A boundary that, when hit, forces the random walk to move back into the permissible state space.
6.1.1.6 7. Explaining the Basic Concept of a Multivariate Autoregressive Model.
When analyzing multiple time series simultaneously, we use multivariate models.
Multivariate Time Series: An
m
-dimensional multivariate time series is a sequence ofm
-dimensional vectors,x_t
, where each vector represents observations ofm
different variables of interest at timet
. These are modeled using sequences of random vectors,X_t
, where the components are denotedX_t(1), ..., X_t(m)
.Vector Autoregressive (VAR) Process: An
m
-dimensionalVAR(p)
process models the current random vectorX_t
as a linear combination ofp
past random vectorsX_{t-1}, ..., X_{t-p}
and anm
-dimensional white noise processe_t
. The defining equation isX_t = μ + A_1(X_{t-1}-μ) + ... + A_p(X_{t-p}-μ) + e_t
, whereA_j
arem x m
matrices.- Properties: A
VAR(p)
process is stationary if all the eigenvalues of its characteristic matrixA
(which isA_1
for aVAR(1)
model) are strictly less than 1 in magnitude. They are always invertible. OnlyVAR(1)
processes possess the Markov property. - Conversion from Univariate AR(p): Interestingly, a univariate
AR(p)
process, which is not Markov forp > 1
, can be rearranged into a multivariateVAR(1)
model, thereby obtaining a Markov property representation. This is achieved by defining a vector of lagged variables.
- Properties: A
6.1.1.7 8. Explaining the Concept of Cointegrated Time Series.
Cointegration describes a specific relationship between non-stationary time series.
- Definition: Two time series processes,
X
andY
, are cointegrated if they meet two key conditions:- Both
X
andY
areI(1)
random processes, meaning they are individually non-stationary but become stationary after being differenced once. - There exists a non-zero linear combination,
αX + βY
, that is stationary. This vector(α, β)
is known as the cointegrating vector.
- Both
- Implication: The core idea is that even if individual series are non-stationary and tend to wander, they do not wander too far from each other, implying a long-term equilibrium relationship between them. This can occur if one process drives the other, or if both are driven by the same underlying process.
I trust this detailed summary, drawing directly from your CS2 study materials, provides a robust foundation for your understanding of time series concepts. Should you have any further questions or require elaboration on specific points, do not hesitate to ask. Keep up the excellent work in your CS2 preparation!