A doubly infinite sequence of random variables is said to be stationary if all of the following conditions hold:
4. is the same for all times .
5. is the same for all times .
6. only depends on .
For a stationary , we can define . Call the ACVF (AutoCovariance Function). Observe that so is a symmetric function of , so we only consider nonnegative .
Define ACF (AutoCorrelation Function):
So .
Note that
Stationarity refers to the model not the data.
Not all time series models are stationary.
ACVF and ACF are only defined for stationary models.
(Gaussian) White Noise Model
. It's easy to check that
2 MA Models
The Moving Average Model (MA) with order is defined by where . Denote as . There are unknown parameters: .
For MA model
(take ). Since is Gaussian white noise, , unless . So we need . Then
It does not depend on , so is stationary, and , and
For , , and
3 Sample ACF
For fixed , the sample ACF at lag is defined as: where
We can simplify by , and (reasonable when is small compared to ). Then define sample ACF:
Note that .
Although sample ACF can be computed for any time series, it is only useful for stationary ones.
Sample ACF is useful in determining : sample ACF after lag is very small/close to 0.
4 Sample PACF
Define sample PACF (Partial AutoCorrelation) of as (estimate of ) when is fit to the data.
Sample PACF is useful in determining in : sample PACF after is very small/close to 0.
Why PACF? Suppose we have data . The correlation is then
Under usual OLS (see here),
Now we also have data on other variables , and dataset becomes . The partial correlation between given is given by : define residual of given as the residual in linear regression and
And for multiple linear regression, denote coefficients (RSS minimizer) as , then
Now for time series setting with with AR(p), we can write
When is stationary,