Stationary processes and autocorrelation are foundational concepts in time series analysis. They give actuaries the tools to model dependencies in data that evolve over time, which directly feeds into pricing, reserving, and risk management decisions.
A stationary process has statistical properties that don't shift as time passes, and autocorrelation measures how observations at different time points relate to each other. Together, these ideas underpin most of the time series models you'll encounter in actuarial work.
Definition of stationary processes
A stationary process is a stochastic process whose statistical properties remain constant over time. The mean, variance, and autocovariance structure don't depend on when you observe the process. This property is what makes stationary processes tractable for modeling: if the rules governing the data don't change, you can use past behavior to forecast the future.
Strict vs weak stationarity
There are two levels of stationarity, and the distinction matters.
Strict stationarity requires that the entire joint probability distribution is invariant under time shifts. Formally, a process is strictly stationary if has the same distribution as for all choices of and any shift . This is a very strong condition because it constrains every moment and every joint relationship, not just the first two.
Weak stationarity (also called second-order or covariance stationarity) is less demanding. It only requires:
- Constant mean: for all
- Time-invariant autocovariance: , depending only on the lag , not on
Strict stationarity implies weak stationarity (assuming finite second moments), but the reverse isn't generally true. The notable exception is Gaussian processes, where weak stationarity does imply strict stationarity, because a Gaussian distribution is fully determined by its first two moments.
In practice, weak stationarity is the version you'll work with most often, since it's both verifiable from data and sufficient for fitting ARMA-type models.
Properties of stationary processes
- The mean is constant over time: for all
- The variance is constant over time: for all
- The autocovariance depends only on the lag:
- The autocorrelation depends only on the lag:
These properties make stationary processes well-suited for modeling phenomena with stable long-term behavior. Think of daily log-returns on a stock index or short-term interest rate changes: the level fluctuates, but the overall statistical character stays roughly the same.
Autocorrelation in stationary processes
Autocorrelation measures the linear dependence between a time series and lagged versions of itself. For actuaries, this is how you quantify whether today's observation tells you something about tomorrow's (or next month's, or next year's).
Definition of autocorrelation
The autocorrelation at lag is defined as:
Here is the autocovariance at lag and is the variance. Dividing by the variance normalizes the measure to fall between and , making it comparable across different series regardless of scale.
Autocorrelation function (ACF)
The autocorrelation function (ACF) plots against the lag . It gives you a visual snapshot of the entire dependence structure of a stationary series. You can use it to:
- Identify how quickly the influence of a shock decays
- Spot seasonal or cyclical patterns (spikes at regular lag intervals)
- Guide model selection (e.g., the ACF of an MA() process cuts off after lag )
Properties of ACF
- Symmetry:
- Normalized at zero: (a series is perfectly correlated with itself)
- Bounded:
- White noise signature: For a white noise process, for all
Sample ACF vs population ACF
The population ACF is a theoretical quantity. In practice, you estimate it from data using the sample ACF:
where is the sample mean and is the number of observations. This estimator is consistent, meaning it converges to the true as . However, for small samples or large lags (where is close to ), the estimate becomes unreliable because fewer pairs of observations contribute to the sum.
Autocovariance in stationary processes
Autocovariance captures the same dependence information as autocorrelation but on the original scale of the data (it's not normalized). It plays a central role in the theoretical development of time series models.
Definition of autocovariance
The autocovariance at lag is:
where is the process mean. Unlike autocorrelation, autocovariance retains the units of the original data (squared), so its magnitude depends on the scale of the series.

Autocovariance function (ACVF)
The autocovariance function (ACVF) plots against the lag . It characterizes the second-order properties of a stationary process and tells you both the direction and magnitude of dependence at each lag.
Properties of ACVF
- Symmetry:
- Variance at lag 0:
- Bounded by variance: for all
- White noise: for all
Relationship between ACF and ACVF
The ACF is simply the ACVF normalized by the variance:
And conversely:
Use the ACF when you want a scale-free measure for comparing dependence across different series. Use the ACVF when you need the actual covariance values, for instance when writing out the variance of a forecast.
Estimation of ACF and ACVF
Estimating these functions from observed data is a critical step in time series analysis. The estimates guide model selection and parameter estimation.
Estimating ACF from data
The sample ACF at lag is:
Note that the denominator sums over all observations (not ). This ensures the sample ACF matrix remains positive semi-definite, which is a necessary property for valid covariance structures.
Estimating ACVF from data
The sample ACVF at lag is:
Again, dividing by (rather than ) preserves positive semi-definiteness. Both the sample ACF and ACVF are consistent estimators of their population counterparts.
Confidence intervals for ACF and ACVF
Under the null hypothesis that the true process is white noise, the sample ACF at each lag is approximately normally distributed with mean 0 and variance . This gives the familiar confidence bands at (for a 95% level) that you see on ACF plots.
For a general stationary process (not white noise), the variance of is more complex. Bartlett's formula gives:
This is harder to use in practice because it depends on the unknown population ACF, but it's important to know that the simple bands are only strictly valid under the white noise assumption.
Models for stationary processes
Several standard models capture different types of dependence in stationary time series. Choosing the right one depends on the ACF and PACF patterns you observe in the data.
Autoregressive (AR) models
An AR() model expresses the current value as a linear combination of its most recent values plus a white noise shock:
where . The process is stationary if all roots of the characteristic polynomial lie outside the unit circle.
The ACF of an AR process decays gradually (exponentially or with damped oscillations), while the partial autocorrelation function (PACF) cuts off after lag . This PACF cutoff is how you identify the order of an AR model from data.

Moving average (MA) models
An MA() model expresses the current value as a linear combination of the current and most recent white noise terms:
MA processes are always stationary (regardless of parameter values). Their ACF cuts off sharply after lag , which is the key identification tool. The PACF, by contrast, decays gradually.
MA models represent processes with "finite memory": a shock at time affects the series for exactly periods and then disappears.
Autoregressive moving average (ARMA) models
An ARMA() model combines both components:
ARMA models are flexible enough to represent a wide range of stationary processes, often with fewer parameters than a pure AR or MA model would require. Both the ACF and PACF of an ARMA process decay gradually, which is what distinguishes it from pure AR or pure MA behavior.
Model selection typically involves examining the sample ACF and PACF, then using information criteria (AIC, BIC) to compare candidate models.
Stationarity tests
Before fitting a stationary model, you need to verify that your data actually is stationary. There are both informal and formal approaches.
Visual inspection of time series
Start by plotting the series. A stationary series should show:
- No visible trend (the mean doesn't drift up or down)
- Roughly constant spread over time (no expanding or contracting variance)
- No obvious structural breaks
This is a quick sanity check, not a definitive test. Visual inspection can miss subtle non-stationarity, and it can also be misleading with short series.
Augmented Dickey-Fuller (ADF) test
The ADF test is the most commonly used formal test for a unit root.
- Null hypothesis: The series has a unit root (non-stationary)
- Alternative hypothesis: The series is stationary
The test regression augments the basic Dickey-Fuller equation with lagged difference terms to account for serial correlation. You reject the null (conclude stationarity) if the test statistic is sufficiently negative, compared to the ADF critical values (which are non-standard and more negative than normal distribution critical values).
Kwiatkowski-Phillips-Schmidt-Shin (KPSS) test
The KPSS test flips the hypotheses:
- Null hypothesis: The series is stationary
- Alternative hypothesis: The series has a unit root (non-stationary)
This reversal is useful because it lets you test stationarity from the opposite direction. A common strategy is to run both the ADF and KPSS tests together. If ADF rejects its null and KPSS fails to reject its null, you have strong evidence of stationarity. If they give conflicting results, further investigation (differencing, transformation) is needed.
Applications of stationary processes
Time series forecasting
Stationary models like AR, MA, and ARMA form the backbone of time series forecasting. Once you've confirmed stationarity and fitted an appropriate model, you can generate point forecasts and prediction intervals for future values. For actuaries, this applies directly to projecting claim frequencies, loss ratios, and economic variables used in reserve calculations.
Signal processing
Spectral analysis and filtering techniques assume stationarity to decompose a time series into frequency components. While this originates in engineering, actuaries encounter these methods when analyzing high-frequency financial data or extracting cyclical patterns from economic indicators.
Quality control
Control charts (Shewhart, CUSUM, EWMA) monitor a process over time and flag deviations from expected behavior. These charts assume the underlying process is stationary, so that deviations represent genuine changes rather than natural drift. Actuaries working in product reliability, warranty analysis, or operational risk may use these tools to detect anomalies.