Skip to main content
Back

Time Series Analysis and Forecasting in Econometrics: Concepts, Methods, and Applications

Study Guide - Smart Notes

Tailored notes based on your materials, expanded with key definitions, examples, and context.

Time Series Data and Forecasting

Prediction vs. Causation

In time series econometrics, it is crucial to distinguish between prediction (forecasting) and causation. Forecasting focuses on predicting future values based on past observations, while causal analysis seeks to understand the effect of one variable on another. For forecasting, the primary goal is accuracy in out-of-sample predictions, not causal interpretation of coefficients.

Types of Time Series Data Transformations

  • Levels: The raw values of a variable over time.

  • Differences (First Differences): The change in the variable from one period to the next, denoted as .

  • Logarithms: The natural log transformation, often used to stabilize variance and interpret growth rates.

  • Percent Changes: The percentage change from one period to the next, often approximated by for small changes.

  • Changes in Logarithms: .

  • Portion of the Total: Expressing a variable as a percentage of a total (e.g., unemployment rate).

Example: Data in First Differences

The first difference transformation is commonly used to analyze changes rather than levels, especially when the data exhibit trends or nonstationarity.

First differences of nonfarm payrolls

Example: Changes in Logarithms, % Change

Taking the difference of logarithms approximates the percentage change, which is useful for interpreting growth rates in economic data.

Percent change in nonfarm payrolls

Example: Consumer Price Index (CPI) Changes

Year-over-year percent changes in the CPI are used to measure inflation.

Percent change in CPI

Lagging the Data and Autocorrelation

Lags and Differences

  • Lag: The value of a variable from a previous period, e.g., the first lag is .

  • First Difference: .

  • First Difference of Logarithm: .

  • Percentage Change Approximation: .

Autocorrelation (Serial Correlation)

Autocorrelation measures the correlation of a time series with its own past values. It is a key concept in time series analysis, indicating the persistence of shocks over time.

  • First Autocovariance:

  • First Autocorrelation:

  • j-th Autocorrelation:

Forecasting Models and Stationarity

Forecasting Basics

Forecasting uses historical data to predict future values. The accuracy of forecasts depends on the similarity between the historical (in-sample) and future (out-of-sample) periods, a property known as stationarity.

Stationarity

A time series is stationary if its probability distribution does not change over time. Stationarity is essential for reliable forecasting, as it ensures that the relationships observed in the past persist into the future.

  • Strict Stationarity: The joint distribution of does not depend on .

  • Joint Stationarity: Two series and are jointly stationary if their joint distribution does not depend on .

Forecasts and Forecast Errors

  • Forecast: An out-of-sample prediction, e.g., is the forecast for based on data through .

  • Forecast Error:

  • Mean Squared Forecast Error (MSFE):

Forecast notation diagram

Autoregressive and ADL Models

Autoregressive (AR) Models

An autoregressive model predicts a variable using its own past values. The order of the model, , indicates how many lags are included.

  • AR(1) Model:

  • AR(p) Model:

Autoregressive Distributed Lag (ADL) Models

The ADL model extends the AR model by including lagged values of other predictors (X variables):

  • ADL(p, q) Model:

Interest rates and term spread

Model Selection and Forecast Evaluation

Information Criteria for Lag Selection

Choosing the appropriate lag length is crucial for model performance. Two common criteria are:

  • Bayesian Information Criterion (BIC):

  • Akaike Information Criterion (AIC):

BIC penalizes model complexity more heavily than AIC, often resulting in more parsimonious models.

Forecast Intervals and Uncertainty

  • Root Mean Squared Forecast Error (RMSFE):

  • Forecast Interval (95%):

Fan chart for forecast intervals

Nonstationarity: Trends and Structural Breaks

Trends in Time Series Data

A trend is a persistent, long-term movement in a time series. Trends can be deterministic (e.g., linear) or stochastic (random walk).

  • Deterministic Trend:

  • Stochastic Trend (Random Walk):

Time series with trendTime series with trend 2Time series with trend 3

Unit Roots and the Dickey-Fuller Test

A unit root in an AR(1) model () implies a stochastic trend. The Dickey-Fuller test is used to test for unit roots:

  • DF Test Regression:

  • Null Hypothesis: (unit root, nonstationary)

  • Alternative Hypothesis: (stationary)

Dickey-Fuller critical values

Structural Breaks and Model Stability

Testing for Structural Breaks

Structural breaks occur when the relationship between variables changes over time. The Chow test (for known break dates) and the Quandt Likelihood Ratio (QLR) test (for unknown break dates) are used to detect such breaks.

  • Chow Test: Tests for parameter stability at a known break date.

  • QLR Test: Maximum of Chow F-statistics over a range of possible break dates.

QLR critical valuesQLR F-statistics over time

Pseudo Out-of-Sample (POOS) Forecasting

POOS forecasting simulates real-time forecasting by re-estimating the model as new data become available and comparing forecasts to actual outcomes. This method is useful for assessing model stability, especially at the end of the sample.

POOS forecasts of GDP growth

Summary Table: Key Concepts and Formulas

Concept

Formula / Description

First Difference

Log Difference (Approx. % Change)

AR(1) Model

ADL(p, q) Model

Forecast Error

MSFE

RMSFE

BIC

AIC

Pearson Logo

Study Prep