Reading: Ch 3 - Shumway and Stoffer
Intro to Autoregressive models¶
Up until now we’ve thought about signals in both the frequency and the time domain, and most recently we’ve been looking at decomposing signals into sinusoids and looking at power spectra. Today, we look at autoregressive models, which allow us to predict a signal from its past.
Autoregression = regression on self.
Ordinary regression:
AR(1) process:
In the AR process, can be written as a function of past values . determines the number of steps in the past needed to forecast the current value. This is also called the order of the AR() process.
AR(p) process:
is stationary
are constants
We can also add an intercept, if .
Let’s look at some examples of these in the accompanying Lecture 16 notebook. In the plot below, series A is white noise, series B is an AR(1) model with , series C is a random walk (AR with , which is not stationary, and series D is an AR(1) model with .

AR(1) process¶
The simplest AR model is the AR(1) model, where depends on plus some white noise (or “shock”). This could be used to look at temperature (if it’s 72 degrees F right now, what’s your best guess for one hour from now?), stock returns (if the market went up 2% today, what does that tell you about tomorrow?), etc.
What does control?
| Behavior | |
|---|---|
| White noise (no memory) | |
| Positive memory - values drift slowly | |
| Very long memory | |
| Random walk (nonstationary) | |
| Oscillatory memory - alternating values |
So what happens if ? In this case, the influence of past shocks doesn’t decay and the variance of increases without bound. The AR(1) process is stationary iff . Below are some example AR1 time series from the accompanying lecture notebook:


Some observations:
As increases, the series becomes smoother - why?
The variance of the series also increases with - why?
At , this looks close to a random walk
Properties of the stationary AR(1) model¶
When , we can derive closed-form expressions (see the derivations of these in your book, Ch 3 example 3.1):
| Property | Formula |
|---|---|
| Mean | (assuming zero-mean) |
| Variance | |
| Autocovariance | for |
| ACF |
Note that the ACF of an AR(1) process decays exponentially! We can also see here that is needed for stationarity, because the denominator in the variance equation goes to zero as . Let’s look at an example for the ACF in the notebook.

The backshift operator¶
Next, to look into the properties of the AR models, we’re going to define the backshift operator :
Then we can rewrite the AR(1) model as:
or the AR(p) model:
The autoregressive operator/characteristic polynomial¶
We then define the autoregressive operator as:
so we have
This is also sometimes called the characteristic polynomial because we can use it to find the roots of the polynomial, which will then tell us some more interpretable information about stationarity and oscillatory frequency. Let’s first consider the AR(1) model. There we have:
The root is at .
Important: For stationarity, we require all roots to lie outside the unit circle, that is, .
Now let’s try this with an AR(2) model:
the characteristic polynomial is:
Now when we solve for the roots of this polynomial, we could potentially get:
Two real roots (overdamped behavior) - this means the process will decay smoothly back to the mean without oscillating
Complex conjugate roots - this will lead to oscillatory decay
Example AR(2) model¶
Let’s try an example:
We will do the following:
Write the characteristic polynomial
Find its roots. Are they real or complex?
Is this process stationary?
What kind of behavior do you expect?
Characteristic polynomial:
We then solve for its roots using the quadratic formula. For this particular AR(2) model, we get complex roots:
The roots are , which are complex. The modulus tells us whether the process is stationary:
Because this is greater than 1, we determine that yes, the process is indeed stationary. The modulus also has another intuitive meaning - a modulus closer to 1 means less damping and a more persistent oscillation, whereas a modulus closer to 0 means more damping (the signal dies out fast).
Next time - ARMA¶
Next time, we will discuss AR(p) models and the extension to ARMA models, which include a moving average comopnent that allows us to also model recent shocks in our autoregressive process.