Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
The STATESPACE Procedure

Preliminary Autoregressive Models

After computing the sample autocovariance matrices, PROC STATESPACE fits a sequence of vector autoregressive models. These preliminary autoregressive models are used to estimate the autoregressive order of the process and limit the order of the autocovariances considered in the state vector selection process.

Yule-Walker Equations for Forward and Backward Models

Unlike a univariate autoregressive model, a multivariate autoregressive model has different forms, depending on whether the present observation is being predicted from the past observations or from the future observations.

Let xt be the r-component stationary time series given by the VAR statement after differencing and subtracting the vector of sample means. (If the NOCENTER option is specified, the mean is not subtracted.) Let n be the number of observations of xt from the input data set.

Let et be a vector white noise sequence with mean vector 0 and variance matrix {{\Sigma}_{p}}, and let nt be a vector white noise sequence with mean vector 0 and variance matrix {{\Omega}_{p}}.Let p be the order of the vector autoregressive model for xt.

The forward autoregressive form based on the past observations is written as follows:

x_{t}=\sum_{i=1}^p{{\Phi}^p_{i}x_{t-i}}+e_{t}

The backward autoregressive form based on the future observations is written as follows:

x_{t}=\sum_{i=1}^p{{\Psi}^p_{i}x_{t+i}}+n_{t}

Letting E denote the expected value operator, the autocovariance sequence for the xt series, {{\Gamma}_{i}}, is

{\Gamma}_{i} = E x_{t} x^{'}_{t-i}

The Yule-Walker equations for the autoregressive model that matches the first p elements of the autocovariance sequence are

[\matrix{
 {\Gamma}_{0} & {\Gamma}_{1} & { ... } & {\Gamma}_{p-1} \cr
 {\Gamma}^...
 ...
[\matrix{
 {\Gamma}_{1} \cr
 {\Gamma}_{2} \cr
 {\vdots} \cr
 {\Gamma}_{p} 
 } ]

and

[\matrix{
 {\Gamma}_{0} &
 {\Gamma}'_{1} &
 { ... } & {\Gamma}'_{p-1} \cr
 {\Gam...
 ...matrix{
 {\Gamma}_{1}' \cr
 {\Gamma}_{2}' \cr
 {\vdots} \cr
 {\Gamma}_{p}' 
 } ]

Here { {\Phi}^p_{i}} are the coefficient matrices for the past observation form of the vector autoregressive model, and { {\Psi}^p_{i}} are the coefficient matrices for the future observation form. More information on the Yule-Walker equations in the multivariate setting can be found in Whittle (1963) and Ansley and Newbold (1979).

The innovation variance matrices for the two forms can be written as follows:

{{\Sigma}}_{p} = {\Gamma}_{0}-
 \sum_{i=1}^p{{\Phi}^p_{i}
 {\Gamma}^{'}_{i}
 }
\vspace{1ex}
{\Omega}_{p} = {\Gamma}_{0} -
 \sum_{i=1}^p{{\Psi}^p_{i}{\Gamma}_{i}
 }

The autoregressive models are fit to the data using the preceding Yule-Walker equations with {{\Gamma}_{i}} replaced by the sample covariance sequence Ci. The covariance matrices are calculated as

C_{i} = \frac{1}{N-1}
 \sum_{t=i+1}^N{x_{t}x_{t-i}^' }

Let {\hat{{\Phi}}_{p}}, {\hat{{\Psi}}_{p}},{\hat{{\Sigma}}_{p}}, and {\hat{{\Omega}}_{p}}represent the Yule-Walker estimates of {{\Phi}_{p}},{{\Psi}_{p}}, {{\Sigma}_{p}}, and {{\Omega}_{p}} respectively. These matrices are written to an output data set when the OUTAR= option is specified.

When the PRINTOUT=LONG option is specified, the sequence of matrices {\hat{{\Sigma}}_{p}}and the corresponding correlation matrices are printed. The sequence of matrices {\hat{{\Sigma}}_{p}}is used to compute Akaike information criteria for selection of the autoregressive order of the process.

Akaike Information Criterion

The Akaike information criterion, or AIC, is defined as -2(maximum of log likelihood)+2(number of parameters). Since the vector autoregressive models are estimates from the Yule-Walker equations, not by maximum likelihood, the exact likelihood values are not available for computing the AIC. However, for the vector autoregressive model the maximum of the log likelihood can be approximated as

{\ln}( L ) {\approx} -\frac{n}2 {\ln}( {|\hat{{\Sigma}}_{p}|} )

Thus, the AIC for the order p model is computed as

AIC_{p} = n {\ln}( {|\hat{{\Sigma}}_{p}|} ) + 2pr^2

You can use the printed AIC array to compute a likelihood ratio test of the autoregressive order. The log-likelihood ratio test statistic for testing the order p model against the order p-1 model is

- n {\ln}( {|\hat{{\Sigma}}_{p}|} )
+ n {\ln}( {|\hat{{\Sigma}}_{p-1}|} )

This quantity is asymptotically distributed as a {{\chi}^2} with r2 degrees of freedom if the series is autoregressive of order p-1. It can be computed from the AIC array as

AICp-1-AICp+2r2

You can evaluate the significance of these test statistics with the PROBCHI function in a SAS DATA step, or with a {{\chi}^2} table.

Determining the Autoregressive Order

Although the autoregressive models can be used for prediction, their primary value is to aid in the selection of a suitable portion of the sample covariance matrix for use in computing canonical correlations. If the multivariate time series xt is of autoregressive order p, then the vector of past values to lag p is considered to contain essentially all the information relevant for prediction of future values of the time series.

By default, PROC STATESPACE selects the order, p, producing the autoregressive model with the smallest AICp. If the value p for the minimum AICp is less than the value of the PASTMIN= option, then p is set to the PASTMIN= value. Alternatively, you can use the ARMAX= and PASTMIN= options to force PROC STATESPACE to use an order you select.

Significance Limits for Partial Autocorrelations

The STATESPACE procedure prints a schematic representation of the partial autocorrelation matrices indicating which partial autocorrelations are significantly greater or significantly less than 0. Figure 18.10 shows an example of this table.

 
The STATESPACE Procedure

Schematic Representation of Partial
Autocorrelations
Name/Lag 1 2 3 4 5 6 7 8 9 10
x ++ +. .. .. .. .. .. .. .. ..
y ++ .. .. .. .. .. .. .. .. ..
+ is > 2*std error,  - is < -2*std error,  . is between
Figure 18.10: Significant Partial Autocorrelations

The partial autocorrelations are from the sample partial autoregressive matrices { \hat{{\Phi}}^p_{p}}.The standard errors used for the significance limits of the partial autocorrelations are computed from the sequence of matrices {{\Sigma}_{p}} and {{\Omega}_{p}}.

Under the assumption that the observed series arises from an autoregressive process of order p-1, the pth sample partial autoregressive matrix { \hat{{\Phi}}^p_{p}}has an asymptotic variance matrix {\frac{1}n {\Omega}^{-1}_{p}{\otimes}{\Sigma}_{p}}.

The significance limits for { \hat{{\Phi}}^p_{p}}used in the schematic plot of the sample partial autoregressive sequence are derived by replacing {{\Omega}_{p}} and {{\Sigma}_{p}}with their sample estimators to produce the variance estimate, as follows:

\hat{Var}( \hat{{\Phi}}^p_{p} ) =
 (\frac{1}{n-rp})
 \hat{{\Omega}}^{-1}_{p}{\otimes}\hat{{\Sigma}}_{p}

Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
Top
Top

Copyright © 1999 by SAS Institute Inc., Cary, NC, USA. All rights reserved.