Wangsheng's World
ARMA Models
Last Update: October 10, 2024
ARMA(\(p, q\)) Processes
  • \(\{X_t\}\) is an ARMA(\(p, q\)) process if \(\{X_t\}\) is stationary solution to \[ \phi(B)X_t = \theta(B)Z_t \] for every \(t\), where \(\{Z_t\} \sim WN(0, \sigma^2)\), and \(\phi(z) = 1 - \phi_1 z - \dotsm - \phi_p z^p\) & \(\theta(z) = 1 + \theta_1 z + \dotsm + \theta_q z^q\) have no common roots.
  • \(\{X_t\}\) is an ARMA(\(p, q\)) process with mean \(\mu\) if \(\{ X_t - \mu \}\) is an ARMA(\(p, q\)) process.
Causality
  • \(\{X_t\}\) is causal if we can write \[ X_t = \sum_{j=0}^\infty \psi_j Z_{t-j} = \psi(B)Z_t, \] where \(\psi(z)\) \( = \sum_{j=0}^\infty \psi_j z^j = \psi_0 + \psi_1 z + \psi_2 z^2 + \dotsm \), with \(\sum_{j=0}^\infty \mid \psi_j \mid < \infty \).
  • \(\psi_j\)'s are called the \(\psi\)-coefficients.
Determining \(\psi\)-coefficients
  • To determine \(\psi_j\)'s,
    • We start with the identity
      \( (1 - \phi_1 z - \dotsm - \phi_p z^p)(\psi_0 + \psi_1 z + \dotsm) = 1 + \theta_1 z + \dotsm + \theta_q z^q \)
    • Equating coefficients of \(z^j, j = 0, 1, \dotsm\), we find that
      \(z^0: \psi_0 = 1\)
      \(z^1: \psi_1 - \phi_1 \psi_0 = \theta_1 \Rightarrow \psi_1 = \theta_1 + \phi_1 \psi_0 = \theta_1 + \phi_1\)
  • In general, \[ \psi_j = \theta_j + \sum_{i=1}^p \phi_i \psi_{j-i} \] for \(j = 0, 1, 2, \dotsm\), with convention \(\theta_0 = 1, \theta_j = 0\) for \(j > q\), and \(\psi_j = 0\) for \(j < 0\).
  • We can solve for \(\psi_0, \psi_1, \psi_2, \dotsm\) recursively.
Examples
  • AR(2): \(X_t - 0.7 X_{t-1} + 0.1 X_{t-2} = Z_t\)
    (Example 3.1.2 from textbook)
    The autoregressive polynomial for this process has the factorization \( \phi(z) = (1 - 0.5z)(1 - 0.2z) \), and is therefore zero at \(z=2\) and \(z=5\). Since these zeros lie outside the unit circle, we conclude that \(\{X_t\}\) is a causal AR(2) process with coefficients \( \{ \psi_j \} \) given by
    • \(\psi_0 = 1\),
    • \(\psi_1 = 0.7 \),
    • \(\psi_2 = 0.7^2 - 0.1\),
    • \(\psi_j = 0.7 \psi_{j-1} - 0.1 \psi_{j-2}, j = 2, 3, \dots\).
    While it is simple matter to calculate \(\psi_j\) numerivally for any \(j\), it is possible also to give an explicit solution of these difference equations using the theory of linear difference equations.
  • ARMA(1, 1) process: \(X_t - 0.5 X_{t-1} = Z_t + 0.4 Z_{t-1}\)
    (see photo - 15:41 Oct 3)
    \(\phi(z) = 1 - 0.5 z \) & \(\theta(z) = 1 + 0.4 z \)
    \( (\psi_0 + \psi_1 z + \psi_2 z^2 + \dotsm)(1-0.5z) = 1 + 0.4z \)
    \( z^0: \psi_0 = 1 \)
    \( z^1: \psi_1 - 0.5\psi_0 = 0.4 \Rightarrow \psi_1 = 0.9 \)
    \( z^2: \psi_2 - 0.5\psi_1 = 0 \Rightarrow \psi_2 = 0.5(0.9) = 0.45 \)
    \( \vdots \)
    \( z^j: \psi_j - 0.5 \psi_{j-1} = 0 \Rightarrow \psi_j = 0.5 \psi_{j-1} \text{ for } j \ge 2 \)
  • R Command: ARMAtoMA
Invertibility
  • \(\{X_t\}\) is invertible if we can write \[ Z_t = \sum_{j=0}^\infty \pi_j X_{t-j} = \pi(B)X_t, \] where \(\pi(z) = \sum_{j=0}^\infty \pi_j z^j = \pi_0 + \pi_1 z + \pi_2 z^2 + \dotsm\), with \(\sum_{j=0}^\infty \mid \pi_j \mid < \infty \).
  • \(\pi_j\)'s are called the \(\pi\)-coefficients.
Determining \(\pi\)-coefficients
  • To determine \(\pi_j\)'s,
    • We start with the identity
      \( (1 + \theta_1 z + \dotsm + \theta_q z^q)(\pi_0 + \pi_1 z + \dotsm) = 1 - \phi_1 z - \dotsm - \phi_p z^p \)
    • Equating coefficients of \(z^j, j = 0, 1, \dotsm\), we find that
      \(z^0: \pi_0 = 1\)
      \(z^1: \pi_1 + \theta_1 \pi_0 = - \phi_1 \Rightarrow \pi_1 = - \phi_1 - \theta_1 \pi_0 = - \phi_1 - \theta_1\)
      \(\vdots\)
  • In general, \[ \pi_j = - \phi_j - \sum_{i=1}^q \theta_i \pi_{j-i} \] for \(j = 0, 1, 2, \dotsm\), with convention \(\phi_0 = -1, \phi_j = 0\) for \(j > p\), and \(\pi_j = 0\) for \(j < 0\).
  • We can solve for \(\pi_0, \pi_1, \pi_2, \dotsm\) recursively.
Examples
  • MA(1): \( X_t = Z_t + \theta Z_{t-1} \)
  • ARMA(1, 1) process: \( X_t - 0.5 X_{t-1} = Z_t + 0.4 Z_{t-1} \)
ACVF and ACF of ARMA(\(p, q\)) Processes
  • MA(\(q\)) processes: \[ X_t = Z_t + \theta_1 Z_{t-1} + \dotsm + \theta_q Z_{t-q} \]
  • AR(\(p\)) processes: \[ X_t - \phi_1 X_{t-1} - \dotsm - \phi_p X_{t-p} = Z_t \]
  • ARMA(\(p, q\)) processes: \[ \begin{align*} & X_t - \phi_1 X_{t-1} - \dotsm - \phi_p X_{t-p} \\ & = Z_t + \theta_1 Z_{t-1} + \dotsm + \theta_q Z_{t-q} \end{align*} \]
  • MA(\(q\)) is a special case of linear processes
  • ACVF is given by
    \( \gamma(h) = \begin{cases} \sigma^2 \sum_{j=0}^{q - \mid h \mid} \theta_j \theta_{j + \mid h \mid} & \text{ if } \mid h \mid \leq q \\ 0 & \text{ if } \mid h \mid > q \end{cases} \)
  • ACF is given by \(\rho(h) = \frac{\gamma(h)}{\gamma(0)}\)
Examples
  • MA(2): \(X_t = Z_t + 0.3 Z_{t-1} - 0.1 Z_{t-2}\)
  • R command: ARMAacf
    • R example: ACF of some MA(\(q\)) processes
Remarks
  • For an MA(\(q\)) process, ACF is 0 for all lags \(\mid h \mid > q\).
  • Converse is also true: if the ACF of a stationary process \(\{X_t\}\) is 0 beyond some lag \(q\), then \(\{X_t\}\) is an MA(\(q\)) process.
    • This suggests that if \(\hat{\rho}(h)\) is small for \(\mid h \mid > q\), then an MA of order \(q\) (or less) may be appropriate.
Example
Consecutive daily overshorts from an underground gasoline tank at a filling station in Colorado.
oshorts.txt
ACVF of AR(\(p\)) Processes - Method 1
  • To calcualte ACVF of a causal AR(\(p\)) process, we can express the process as a linear process, i.e.,
    \( X_t = \sum_{j=0}^\infty \psi_j Z_{t-j}, \{Z_t\} \sim WN(0, \sigma^2) \)
  • Then, the ACVF is readily calculated by
    \( \gamma(h) = \sigma^2 \sum_{j=0}^\infty \psi_j \psi_{j + \mid h \mid} \).
ACVF of AR(\(p\)) Processes - Method 2
(convenient for finding closed-form solution)
  • Obtain a homogeneous difference equation
    \( \gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h - p) = 0 \), \(h \ge p\)
  • Let \(\xi_1, \dotsm, \xi_k\) denote the distinct roots of \(\phi(z)\), each with multiplicity \(r_1, \dotsm, r_k\).
    \( r_1 + \dotsm + r_k = p\)
  • Then, \(\gamma(h) = \xi_1^{-h}P_1(h) + \dotsm + \xi_k^{-h}P_k(h)\) for \(h \ge 0\), where \(P_i(h)\) is a polynomial in \(h\) of degree \(r_i - 1\).
  • So, \(\gamma(h)\) decays exponentially as \(h \rightarrow \infty\).
    (short-memory processes)
  • \(P_i(h), i = 1, \dotsm, k\), are determined uniquely from initial condictions:
    • \(\gamma(0) - \phi_1 \gamma(1) - \dotsm - \phi_p \gamma(p) = \sigma^2\)
    • \(\gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h-p) = 0, 1 \le h < p\)
  • Examples:
    • AR(2): \(X_t - 0.7X_{t-1} + 0.1 X_{t-2} = Z_t\), \(\{Z_t\} \sim WN(0, \sigma^2)\)
      \( \phi(z) = 1 - 0.7 z + 0.1 z^2 = (1-0.2z)(1-0.5z) \)
      So, roots are \(\xi_1 = 5\) and \(\xi_2 = 2\)
      The general solution is \(\gamma(h) = \xi_1^{-h}P_1(h) + \xi_2^{-h} P_2(h)\) for \(h \ge 0\)
      \( \gamma(h) = 5^{-h} \beta_{10} + 2^{-h} \beta_{20} \), where \(\beta_{10} \) and \(\beta_{20}\) are determined by
      \( \begin{cases} \gamma(0) - 0.7 \gamma(1) + 0.1 \gamma(2) = \sigma^2 \\ \gamma(1) - 0.7 \gamma(0) + 0.1 \gamma(1) = 0 \end{cases} \)
      note that \(\gamma(0) = \beta_{10} + \beta_{20}\), \(\gamma(1) = \frac{\beta_{10}}{5} + \frac{\beta_{20}}{2}\), and \(\gamma(2) = \frac{\beta_{10}}{25} + \frac{\beta_{20}}{4}\)
      We then have \( \begin{cases} 0.864 \beta_{10} + 0.675 \beta_{20} = \sigma^2 \\ -0.48 \beta{10} - 0.15 \beta_{20} = 0 \end{cases} \)
      \(\Rightarrow \beta_{10} = - 0.7716 \sigma^2 \) & \(\beta_{20} = 2.4691 \sigma^2 \)
      Therefore, \(\gamma(h) = -0.7716 \sigma^2 5^{-h} + 2.4691 \sigma^2 2^{-h}\) for \(h \ge 0\).
    • R Example: ACF of some AR(2) processes
      (see photo - 15:45 Oct 8)
ACVF of AR(\(p\)) Processes - Method 3
(convenient for finding numerical solution)
  • We obtain Yule-Walker equations:
    • \(\gamma(0) - \phi_1 \gamma(1) - \dotsm - \phi_p \gamma(p) = \sigma^2\)
    • \(\gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h-p) = 0, h \ge 1\)
  • (1) Solve the first \(p + 1\) equations for \(\gamma(0), \dots, \gamma(p)\)
  • (2) Then, use the subsequent equations to solve for \(\gamma(p + 1), \dots\), successively.
  • Example:
    • AR(2): \(X_t - 0.7 X_{t-1} + 0.1 X_{t-2} = Z_t\), \(\{Z_t\} \sim WN(0, \sigma^2)\)
      (1) Find \(\gamma(0)\), \(\gamma(1)\), and \(\gamma(2)\) using
      \( \begin{cases} \gamma(0) - 0.7 \gamma(1) + 0.1 \gamma(2) = \sigma^2 \\ \gamma(1) - 0.7 \gamma(0) + 0.1 \gamma(1) = 0 \\ \gamma(2) - 0.7 \gamma(1) + 0.1 \gamma(0) = 0 \end{cases} \)
      (2) Find \(\gamma(h), h \ge 3 \) using \(\gamma(h) = 0.7 \gamma(h-1) - 0.1 \gamma(h-2)\)
ACVF of ARMA(\(p, q\)) Processes - Method 1
  • To calculate ACVF of a causal ARMA(\(p, q\)) process, we can express the process as a linear process, i.e.,
    \( X_t = \sum_{j=0}^\infty \psi_j Z_{t-j}, \{Z_t\} \sim WN(0, \sigma^2) \)
  • Then, the ACVF is readily calcualted by
    \( \gamma(h) = \sigma^2 \sum_{j=0}^\infty \psi_j \psi_{j + \mid h \mid} \).
ACVF of ARMA(\(p, q\)) Processes - Method 2
  • Obtain a homogeneous difference equation
    \( \gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h-p) = 0, h \ge m \)
    where \(m = \max(p, q+1)\).
  • Let \(\xi_1, \dotsm, \xi_k\) denote the distinct roots of \(\phi(z)\), each with multiplicity \(r_1, \dotsm, r_k\).
    • \(r_1 + \dotsm + r_k = p\)
  • Then, \(\gamma(h) = \xi_1^{-h}P_1(h) + \dotsm + \xi_k^{-h}P_k(h)\) for \(h \ge m-p\), where \(P_i(h)\) is a polynomial in \(h\) of degree \(r_i - 1\).
    • So, \(\gamma(h)\) decays exponentially as \(h \rightarrow \infty\).
  • \(P_i(h), i = 1, \dotsm, k\), and \(\gamma(j), 0 \leq j < m-p\), are determined uniquely from initial conditions: \[ \begin{align*} & \gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h-p) \\ & = \sigma^2 \sum_{j=0}^{q-h} \theta_{h+j} \psi_j, 0 \leq h < m \end{align*} \]
  • Example:
    ARMA(1, 1): \(X_t - 0.5X_{t-1} = Z_t + 0.4 Z_{t-1}\), \(\{Z_t\} \sim WN(0, \sigma^2)\)
    (easy to check it's causal and invertible)
    \(p = 1, q = 1, \text{ and } m = 2\)
    (1) \( \phi(z) = 1 - 0.5z \Rightarrow \text{ root } \xi = 2 \) with \(r = 1\)
    (2) General solution: \(\gamma(h) = 2^{-h} \cdot \beta_{10}\) for \(h \ge m-p =1\)
    where \(\gamma(0)\) and \(\beta_{10}\) are determined by
    \(\gamma(0) - 0.5 \gamma(1) = 1.36 \sigma^2\)
    \(\gamma(1) - 0.5 \gamma(0) = 0.4 \sigma^2\)
    or
    \(\gamma(0) - 0.25 \beta_{10} = 1.36 \sigma^2\)
    \(0.5 \beta_{10} - 0.5 \gamma(0) = 0.4 \sigma^2\)
    We have \(\beta_{10} = 2.88 \sigma^2\) and \(\gamma(0) = 2.08 \sigma^2\)
    Thus, \(\gamma(0) = 2.08 \sigma^2\) and \(\gamma(h) = 2.88 \sigma^2 \cdot 2^{-h}\) for \(h \ge1\).
    Note:
    rhs \(= \sigma^2 \sum_{j=0}^1 \theta_j \psi_j = \sigma^2(\psi_0 + \theta_1 \psi_1)\) (when \(h = 0\)) (\(\theta_1 = 0.4\), \(\psi_1 = 0.9\), \(\psi_0 = 1\))
    rhs \(= \sigma^2 \theta_1\) (when \(h=1)\)
ACVF of ARMA(\(p, q\)) Processes - Method 3
  • Numerical determination of ACVF from equations
    • \( \gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h-p) \)
      \( = \sigma^2 \sum_{j=0}^{q-h} \theta_{h+j} \psi_j, \text{ for } 0 \leq h < m \)
    • \( \gamma(h) - \phi_1 \gamma(h-1) - \dotsm - \phi_p \gamma(h-p) = 0, \text{ for } h \ge m = \max(p, q+1)\)
    can be carried out by
    • finding \(\gamma(0), \dotsm, \gamma(p)\) from the first \(p + 1\) equations
    • finding \(\gamma(p+1), \dotsm \), successively from subsequent equations.
  • Example:
    ARMA(1, 1): \( X_t - \phi X_{t-1} = Z_t + \theta Z_{t-1}, \mid \phi \mid < 1 \) and \(\{Z_t\} \sim WN(0, \sigma^2) \)
    (general solution in textbook)
    Here consider ARMA(1, 1): \(X_t - 0.5X_{t-1} = Z_t + 0.4 Z_{t-1}\), \(\{Z_t\} \sim WN(0, \sigma^2)\)
    (1) Find \(\gamma(0)\) and \(\gamma(1)\) using
    \(\gamma(0) - 0.5 \gamma(1) = 1.36 \sigma^2\)
    \(\gamma(1) - 0.5 \gamma(0) = 0.4 \sigma^2\)
    \(\Rightarrow \gamma(0) = \dotsm\) and \(\gamma(1) = \dotsm\).
    (2) Find \(\gamma(h), h \ge 2\) using
    \(\gamma(h) - 0.5 \gamma(h-1) = 0 \Rightarrow \gamma(h) = 0.5 \gamma(h-1)\)
    We then have \(\gamma(h) = 0.5^{h-1} \gamma(1)\)
Characteristics of ACVF/ACF of MA, AR, & ARMA
  • MA(\(q\)): Cuts off at lag \(q\);
    i.e., \(= 0\) for \(h > q\).
  • AR(\(p\)): Decreases geometrically as \(h\) increases.
  • ARMA(\(p ,q\)): Decreases geometrically as \(h\) increases.
Partial Autocorrelation Function (PACF) of ARMA(\(p, q\)) Processes
  • MA(\(q\)) model can be identified from its ACF: non-zero up to lag \(q\) and zero afterwards.
  • We need a similar tool for identifying AR(\(p\)) model;
    this role can be filled by PACF.
PACF
  • The PACF of a stationary time series \(\{X_t\}\) at time lag \(h\) denoted as \(\alpha(h)\), is defined as \[ \begin{align*} \alpha(0) & = 1 \\ \alpha(h) & = \phi_{hh}, h \ge 1, \end{align*} \] where \(\phi_{hh}\) is the last component of \(\boldsymbol{\phi}_h = \boldsymbol{\Gamma}_h^{-1} \boldsymbol{\gamma}_h\), with \(\boldsymbol{\phi}_h = (\phi_{h1}, \dotsm, \phi_{hh})'\), \(\boldsymbol{\Gamma}_h = [\gamma(i - j)]_{i,j = 1}^h\), and \( \boldsymbol{\gamma}_h = (\gamma(1), \dotsm, \gamma(h))' \).
  • From Chapter 2, the one-step BLP
    \( P_h X_{h+1} = \mu + \phi_{h1}(X_h - \mu) + \dotsm + \phi_{hh}(X_1 - \mu) \)
    where \(\boldsymbol{\phi}_h\) is determined by \(\boldsymbol{\Gamma}_h \boldsymbol{\phi}_h = \boldsymbol{\gamma}_h\).
  • So, \(\alpha(h)\) is equal to the coefficient associated with \(X_1\) in the expression of \(P_h X_{h+1}\).
    • Consequently, \(\alpha(h)\) can be computed recursively using Durbin-Levinson algorithm.
Examples
  • AR(\(p\)) process: \(X_t = \phi_1 X_{t-1} + \dotsm + \phi_p X_{t-p} + Z_t \), \(\{Z_t\} \sim WN(0, \sigma^2)\)
    From Chapter 2, \( P_n X_{n+1} = \phi_1 X_n + \dotsm + \phi_p X_{n+1-p} \) for \(n \ge p\)
    • For \(h = p\), \(P_h X_{n+1} = \phi_1 X_h + \dotsm + \phi_p X_{h+1-p} = \phi_1 X_p + \dotsm + \phi_p X_1\)
      Then, \(\alpha(p) = \phi_p\)
    • For \(h > p\), \(P_h X_{n+1} = \phi_1 X_h + \dotsm + \phi_p X_{h+1-p} + \dotsm + 0 \cdot X_1\)
      That is \(\alpha(h) = 0\).
    • For \(h < p\), \(\alpha(h) \) can be found from definition.
  • MA(1) process: \(X_t = Z_t + \theta Z_{t-1}, \mid \theta \mid < 1\) & \(\{Z_t\} \sim WN(0, \sigma^2)\)
    (Problem 3.12, see photo - 15:39 Oct 10)
    PACF for MA(1) decays exponentially. So do PACF's for MA(\(q\)) and ARMA(\(p, q\)).
R Examples
  • R command: ARMAacf with option 'pacf = TRUE'
  • PACF of some AR(\(p\)) processes
  • PACF of some MA(\(q\)) processes
Characteristics of ACF & PACF of MA, AR, & ARMA
  • ACF:
    • MA(\(q\)): Cuts off at lag \(q\); i.e., \(= 0\) for \(h > q\).
    • AR(\(p\)): Decreases geometrically as \(h\) increases.
    • ARMA(\(p, q\)): Decreases geometrically as \(h\) increases.
  • PACF:
    • MA(\(q\)): Decreases geometrically as \(h\) increases.
    • AR(\(p\)): Cuts off at lag \(p\); i.e., \(= 0\) for \(h > p\).
    • ARMA(\(p, q\)): Decreases geometrically as \(h\) increases.
Sample PACF
  • Given observations \(x_1, \dotsm, x_n\)
  • Suppose \(x_i \ne x_j\) for some \( i \ne j \).
  • The sample PACF at time lag \(h\), denoted as \(\hat{\alpha}(h)\), is defined as \[ \begin{align*} \hat{\alpha}(0) & = 1 \\ \hat{\alpha}(h) & = \hat{\phi}_{hh}, 1 \leq h < n, \end{align*} \] where \(\hat{\phi}_{hh}\) is the last component of \(\hat{\boldsymbol{\phi}}_h = \hat{\boldsymbol{\Gamma}}_h^{-1} \hat{\boldsymbol{\gamma}}_h \), with \(\hat{\boldsymbol{\phi}}_h = (\hat{\phi}_{h1}, \dotsm, \hat{\phi}_{hh})'\), \( \hat{\boldsymbol{\Gamma}}_h = [\hat{\gamma}(i - j)]_{i, j = 1}^h \), and \( \boldsymbol{\gamma}_h = (\hat{\gamma}(1), \dotsm, \hat{\gamma}(h) )' \).
Remarks
  • \(\hat{\alpha}(h)\) is expected to reflect the properties of \(\alpha(h)\).
  • \(\hat{\alpha}(h)\) can be computed recursively using Durbin-Levinson algorithm.
Sample PACF of AR(\(p\)) Processes
  • If \(X_1, \dotsm, X_n\) is a realization of an AR(\(p\)) process \(\{X_t\}\) with \(\{Z_t\} \sim \text{ iid } (0, \sigma^2)\),
  • then, for large \(n\), \(\hat{\alpha}(\cdot)\)'s at lag greater than \(p\) are approximately independent \(\mathcal{N}(0, 1/n)\) random variables.
  • So, a simple method to identify AR order is:
    • If \(\mid \hat{\alpha}(k) \mid \leq 1.96 / \sqrt{n}\) for \(k > p\), then, an AR of order \(p\) (or less) may be appropriate.
R Example
  • R command: acf with option 'type = partial'
  • The sunspot numbers for the years 1770-1869;
    sunspot.txt
Forecasting ARMA(\(p, q\)) Processes
  • Consider a causal ARMA(\(p, q\)) process
    \( \phi(B) X_t = \theta(B) Z_t, \{Z_t\} \sim WN(0, \sigma^2) \).
  • We assume \(p \ge 1\) and \(q \ge 1\) without loss of generality.
One-Step Prediction of ARMA(\(p, q\))
  • Let \(m = \max(p, q)\)
  • Define \( W_t = \begin{cases} \sigma^{-1} X_t & t = 1, \dotsm, m \\ \sigma^{-1} \phi(B) X_t & t > m \end{cases} \)
  • We apply the innovations algorithm to \(\{W_t\}\), the transformed process.
  • By doing so, a drastic simplification may be achieved in the calculations.
  • Applying the innovations algorithm to \(\{W_t\}\) yields
    \( \hat{W}_{n+1} = \begin{cases} \sum_{j=1}^n \theta_{nj}(W_{n+1-j} - \hat{W}_{n+1-j}) & 1 \leq n < m \\ \sum_{j=1}^q \theta_{nj}(W_{n+1-j} - \hat{W}_{n+1-j}) & n \ge m \end{cases} \)
    where \(\theta_{nj}\) and \(r_n\)\( = \mathbb{E}(W_{n+1} - \hat{W}_{n+1})^2 \) are found recursively, where \(\hat{W}_1 = 0\) and \(\hat{W}_k = P_{k-1}W_k\) for \(k > 1\).
  • Finally, we find \(\hat{X}_{n+1} = P_n X_{n+1} \) from \(\hat{W}_{n+1}\) as follows:
    • \( \hat{X}_{n+1} = \sum_{j=1}^n \theta_{nj} (X_{n+1-j} - \hat{X}_{n+1-j}), 1 \leq n < m \)
    • \( \hat{X}_{n+1} = \phi_1 X_n + \dotsm + \phi_p X_{n+1-p} + \sum_{j=1}^q \theta_{nj} (X_{n+1-j} - \hat{X}_{n+1-j}), n \ge m \)
  • And \(\mathbb{E}(X_{n+1} - \hat{X}_{n+1})^2 = \sigma^2 r_n \)
    • It can be shown that \(X_t - \hat{X}_t = \sigma(W_t - \hat{W}_t) \) for all \(t \ge 1\).
  • The one-step predictors \(\hat{X}_2, \hat{X}_3, \dotsm\) are determined recursively.
Remarks
  • \(\theta_{nj}\) vanishes when \(n \ge m\) and \(j > q\).
  • \(\theta_{nj}\) and \(r_n\) depend only on \(\phi_1, \dotsm, \phi_p\) and \(\theta_1, \dotsm, \theta_q\) but not on \(\sigma^2\).
  • If \(\{X_t\}\) is invertible, then \(\theta_{nj} \rightarrow \theta_j, j = 1, \dotsm, q\), and \(r_n \rightarrow 1\) as \(n \rightarrow \infty\).
Examples
  • Prediction of AR(\(p\)) process: for \(n \ge p\),
    \( P_n X_{n+1} = \phi_1 X_n + \dotsm + \phi_p X_{n+1-p} \)
  • Prediction of MA(\(q\)) process: for \( n \ge 1\),
    \( \hat{X}_{n+1} = \sum_{j=1}^{\min(n,q)} \theta_{nj} (X_{n+1-j} - \hat{X}_{n+1-j}) \)
    where \(\theta_{nj}\)'s are found by applying innovations algorithm to \(\{W_t\}\).
    • In this case, \(W_t = \sigma^{-1}X_t\) for all \(t\).
  • Prediction of ARMA(1, 1) process,
    \( X_t - \phi X_{t-1} = Z_t + \theta Z_{t-1} \) where \(\{Z_t\} \sim WN(0, \sigma^2)\) and \(\mid \phi \mid < 1\):
    \( \hat{X}_{n+1} = \phi X_n + \theta_{n1}(X_n - \hat{X}_n) \)
    for \(n \ge 1\), where \(\theta_{n1}\) is found by applying innovations algorithm to \(\{W_t\}\).
    • In this case,
      \(r_0 = \frac{1 + 2 \theta \phi + \theta^2}{1 - \phi^2}\); \(\theta_{n1} = \frac{\theta}{r_{n-1}}\); \( r_n = 1 + \theta^2 - \frac{\theta^2}{r_{n-1}}\).