Time series analysis often relies on the idea of stationarity, which describes how the statistical behavior of data changes—or does not change—over time. There are two commonly used notions of stationarity: strong (strict) stationarity and weak (wide-sense) stationarity. Understanding the difference between them helps clarify how time series models work and why some models are easier to use than others.
1. Strong (Strict) Stationarity
What strong stationarity means
Consider a stochastic process , which is a sequence of random variables indexed by time.
This process is called strongly stationary if its entire probabilistic structure is unchanged by shifting time.
More precisely:
- For any collection of time points ,
- and for any time shift ,
- the joint distribution of
is exactly the same as the joint distribution of
.
In simple terms:
The probabilistic behavior of the process looks exactly the same, no matter when you observe it.
Consequences of strong stationarity
If a process is strongly stationary, then automatically:
- All have the same distribution
- The mean and variance do not depend on time
- Any dependence structure between time points depends only on how far apart they are, not on their absolute positions in time
Why this definition is demanding
Strong stationarity requires knowledge of all joint distributions of all orders. In real applications, this level of information is rarely available, which makes strong stationarity difficult to verify or use directly.
2. Covariance and Dependence Over Time
Before introducing weak stationarity, it helps to review how dependence between random variables is measured.
Covariance and correlation
For two random variables and :
- Covariance measures linear dependence:
- Variance is a special case:
- Correlation is a normalized version of covariance: and always lies between −1 and 1.
If two variables are independent, both covariance and correlation are zero.
3. The Covariance Function of a Process
To describe how a time series depends on itself across time, we use the covariance function.
For a process :
This function tells us how strongly values at time s and time t are linearly related.
Examples
White noise
- Mean = 0
- Variance = constant
- No correlation between different time points
Covariance structure:
- for
White noise represents pure randomness with no memory.
Random walk
A random walk is formed by summing white noise terms over time.
Key properties:
- Mean stays at zero
- Variance grows with time
- Covariance depends on the time index itself
This dependence on absolute time means a random walk is not stationary.
4. Weak (Wide-Sense) Stationarity
Motivation
Strong stationarity is often too strict. A weaker and much more practical concept focuses only on means and covariances, rather than full distributions.
Definition of weak stationarity
A process is weakly stationary if:
- Constant mean
- Time-invariant covariance structure depends only on the lag , not on the specific time .
This means:
The average level is stable over time, and the way observations relate across time depends only on how far apart they are.
Relationship to strong stationarity
- Strong stationarity ⇒ weak stationarity (if moments exist)
- Weak stationarity does not require full distributional invariance
Because it is much easier to verify and still very useful, weak stationarity is the standard assumption in most practical time series modeling.
5. Autocovariance and Autocorrelation
For a stationary process, the covariance function simplifies.
Autocovariance function (ACVF)
Key properties:
- equals the variance
Autocorrelation function (ACF)
Properties:
- Values lie between −1 and 1
The ACF shows how strongly a time series is correlated with its past values at different lags.
6. Examples of Stationary and Non-Stationary Processes
White noise
- Weakly stationary
- Zero autocorrelation at all non-zero lags
Random walk
- Not weakly stationary
- Variance and covariance grow over time
7. Moving Average Process: MA(1)
A simple but important stationary model is the MA(1) process:
where:
- is white noise
- is a constant
Key results:
- Mean is constant
- Variance is constant
- Covariance is non-zero only at lags 0 and ±1
This makes MA(1) weakly stationary and easy to analyze.
8. Sample Versions: Working with Real Data
In practice, we do not observe the underlying process—only a finite dataset:
From this data, we compute:
- Sample mean
- Sample autocovariance
- Sample autocorrelation
These sample quantities estimate their theoretical counterparts and are meaningful primarily when the data plausibly come from a stationary process.
9. Correlograms (ACF Plots)
A correlogram is a plot of the sample autocorrelation function.
Why it is useful:
- Helps identify whether observations are correlated
- Provides clues about suitable models
- Indicates whether data resemble white noise or structured dependence
Interpreting the confidence bands
Typical ACF plots include horizontal dashed lines around zero. These indicate approximate bounds for random variation when no true correlation exists.
If most points lie within these bounds:
- The series is consistent with having little or no autocorrelation
If clear patterns appear (e.g., significant spikes at certain lags):
- The data likely contain temporal dependence
10. Practical Takeaway
- Strong stationarity is a mathematically complete but rarely practical concept
- Weak stationarity focuses on stable mean and covariance and is widely used
- Autocovariance and autocorrelation summarize temporal dependence
- Sample ACF plots are essential tools for understanding real time series data
In short:
Weak stationarity provides a realistic balance between mathematical structure and practical usability, making it the foundation of most time series analysis.
