Time Series analysis tsa
¶
statsmodels.tsa
contains model classes and functions that are useful
for time series analysis. Basic models include univariate autoregressive models (AR),
vector autoregressive models (VAR) and univariate autoregressive moving average models
(ARMA). Non-linear models include Markov switching dynamic regression and
autoregression. It also includes descriptive statistics for time series, for example autocorrelation, partial
autocorrelation function and periodogram, as well as the corresponding theoretical properties
of ARMA or related processes. It also includes methods to work with autoregressive and
moving average lag-polynomials.
Additionally, related statistical tests and some useful helper functions are available.
Estimation is either done by exact or conditional Maximum Likelihood or conditional least-squares, either using Kalman Filter or direct filters.
Currently, functions and classes have to be imported from the corresponding module, but the main classes will be made available in the statsmodels.tsa namespace. The module structure is within statsmodels.tsa is
stattools : empirical properties and tests, acf, pacf, granger-causality, adf unit root test, kpss test, bds test, ljung-box test and others.
ar_model : univariate autoregressive process, estimation with conditional and exact maximum likelihood and conditional least-squares
arima.model : univariate ARIMA process, estimation with alternative methods
statespace : Comprehensive statespace model specification and estimation. See the statespace documentation.
vector_ar, var : vector autoregressive process (VAR) and vector error correction models, estimation, impulse response analysis, forecast error variance decompositions, and data visualization tools. See the vector_ar documentation.
arma_process : properties of arma processes with given parameters, this includes tools to convert between ARMA, MA and AR representation as well as acf, pacf, spectral density, impulse response function and similar
sandbox.tsa.fftarma : similar to arma_process but working in frequency domain
tsatools : additional helper functions, to create arrays of lagged variables, construct regressors for trend, detrend and similar.
filters : helper function for filtering time series
regime_switching : Markov switching dynamic regression and autoregression models
Some additional functions that are also useful for time series analysis are in other parts of statsmodels, for example additional statistical tests.
Some related functions are also available in matplotlib, nitime, and scikits.talkbox. Those functions are designed more for the use in signal processing where longer time series are available and work more often in the frequency domain.
Descriptive Statistics and Tests¶
|
Estimate autocovariances. |
|
Calculate the autocorrelation function. |
|
Partial autocorrelation estimate. |
|
Partial autocorrelation estimated with non-recursive yule_walker. |
|
Calculate partial autocorrelations via OLS. |
|
Calculate Burg"s partial autocorrelation estimator. |
|
Calculate the crosscovariance between two series. |
|
The cross-correlation function. |
|
Augmented Dickey-Fuller unit root test. |
|
Kwiatkowski-Phillips-Schmidt-Shin test for stationarity. |
|
Range unit-root test for stationarity. |
Zivot-Andrews structural-break unit-root test. |
|
|
Test for no-cointegration of a univariate equation. |
|
BDS Test Statistic for Independence of a Time Series |
|
Compute Ljung-Box Q Statistic. |
Test for heteroskedasticity of residuals |
|
|
Four tests for granger non causality of 2 time series. |
|
Levinson-Durbin recursion for autoregressive processes. |
|
Innovations algorithm to convert autocovariances to MA parameters. |
|
Filter observations using the innovations algorithm. |
|
Levinson-Durbin algorithm that returns the acf and ar coefficients. |
|
Compute information criteria for many ARMA models. |
|
Perform automatic seasonal ARIMA order identification using x12/x13 ARIMA. |
|
Perform x13-arima analysis for monthly or quarterly data. |
Estimation¶
The following are the main estimation classes, which can be accessed through statsmodels.tsa.api and their result classes
Univariate Autoregressive Processes (AR)¶
The basic autoregressive model in Statsmodels is:
|
Autoregressive AR-X(p) model |
|
Class to hold results from fitting an AutoReg model. |
|
Autoregressive AR-X(p) model order selection. |
The ar_model.AutoReg model estimates parameters using conditional MLE (OLS), and supports exogenous regressors (an AR-X model) and seasonal effects.
AR-X and related models can also be fitted with the arima.ARIMA class and the SARIMAX class (using full MLE via the Kalman Filter).
Autoregressive Moving-Average Processes (ARMA) and Kalman Filter¶
Basic ARIMA model and results classes are as follows:
|
Autoregressive Integrated Moving Average (ARIMA) model, and extensions |
|
Class to hold results from fitting an SARIMAX model. |
This model allows estimating parameters by various methods (including conditional MLE via the Hannan-Rissanen method and full MLE via the Kalman filter). It is a special case of the SARIMAX model, and it includes a large number of inherited features from the state space models (including prediction / forecasting, residual diagnostics, simulation and impulse responses, etc.).
Exponential Smoothing¶
Linear and non-linear exponential smoothing models are available:
|
Holt Winter's Exponential Smoothing |
|
Simple Exponential Smoothing |
|
Holt's Exponential Smoothing |
|
Results from fitting Exponential Smoothing models. |
Separately, linear and non-linear exponential smoothing models have also been implemented based on the “innovations” state space approach. In addition to the usual support for parameter fitting, in-sample prediction, and out-of-sample forecasting, these models also support prediction intervals, simulation, and more.
|
ETS models. |
|
Results from an error, trend, seasonal (ETS) exponential smoothing model |
Finally, linear exponential smoothing models have also been separately implemented as a special case of the general state space framework (this is separate from the “innovations” state space approach described above). Although this approach does not allow for the non-linear (multiplicative) exponential smoothing models, it includes all features of state space models (including prediction / forecasting, residual diagnostics, simulation and impulse responses, etc.).
|
Linear exponential smoothing models |
|
Results from fitting a linear exponential smoothing model |
ARMA Process¶
The following are tools to work with the theoretical properties of an ARMA process for given lag-polynomials.
|
Theoretical properties of an ARMA process for specified lag-polynomials. |
|
Find arma approximation to ar process. |
|
A finite-lag AR approximation of an ARMA process. |
|
A finite-lag approximate MA representation of an ARMA process. |
|
Theoretical autocorrelation function of an ARMA process. |
|
Theoretical autocovariances of stationary ARMA processes |
|
Simulate data from an ARMA. |
Compute the impulse response function (MA representation) for ARMA process. |
|
|
Theoretical partial autocorrelation function of an ARMA process. |
|
Periodogram for ARMA process given by lag-polynomials ar and ma. |
|
Deconvolves divisor out of signal, division of polynomials for n terms |
|
Expand coefficients to lag poly |
Remove zeros from lag polynomial |
|
|
AR representation of fractional integration |
|
MA representation of fractional integration |
return coefficients for seasonal difference (1-L^s) |
|
fft tools for arma processes |
Autoregressive Distributed Lag (ARDL) Models¶
Autoregressive Distributed Lag models span the space between
autoregressive models (AutoReg
)
and vector autoregressive models (VAR
).
|
Autoregressive Distributed Lag (ARDL) Model |
|
Class to hold results from fitting an ARDL model. |
|
ARDL order selection |
|
Results from an ARDL order selection |
The ardl.ARDL model estimates parameters using conditional MLE (OLS)
and allows for both simple deterministic terms (trends and seasonal
dummies) as well as complex deterministics using a
DeterministicProcess
.
AR-X and related models can also be fitted with
SARIMAX
class (using full MLE via
the Kalman Filter).
Error Correction Models (ECM)¶
Error correction models are reparameterizations of ARDL models that regress the difference of the endogenous variable on the lagged levels of the endogenous variables and optional lagged differences of the exogenous variables.
|
Unconstrained Error Correlation Model(UECM) |
|
Class to hold results from fitting an UECM model. |
|
|
Statespace Models¶
See the statespace documentation.
Vector ARs and Vector Error Correction Models¶
See the vector_ar documentation.
Regime switching models¶
|
First-order k-regime Markov switching regression model |
|
Markov switching regression model |
Time Series Filters¶
|
Filter a time series using the Baxter-King bandpass filter. |
|
Hodrick-Prescott filter. |
|
Christiano Fitzgerald asymmetric, random walk filter. |
|
Linear filtering via convolution. |
|
Autoregressive, or recursive, filtering. |
|
Filter multiple time series into a single time series. |
|
Convolve two N-dimensional arrays using FFT. |
|
Convolve two N-dimensional arrays using FFT. |
|
Seasonal decomposition using moving averages. |
|
Season-Trend decomposition using LOESS. |
|
Results class for seasonal decompositions |
TSA Tools¶
|
Returns an array with lags included given an array. |
|
Add a trend and/or constant to an array. |
|
Detrend an array with a trend of given order along axis 0 or 1. |
|
Create 2d array of lags. |
|
Generate lagmatrix for 2d array, columns arranged by variables. |
VARMA Process¶
|
class to keep track of Varma polynomial format |
Interpolation¶
|
Modified Denton's method to convert low-frequency to high-frequency data. |
Deterministic Processes¶
Deterministic processes simplify creating deterministic sequences with time
trend or seasonal patterns. They also provide methods to simplify generating
deterministic terms for out-of-sample forecasting. A
DeterministicProcess
can be directly
used with AutoReg
to construct complex
deterministic dynamics and to forecast without constructing exogenous trends.
|
Container class for deterministic terms. |
|
Constant and time trend determinstic terms |
|
Seasonal dummy deterministic terms |
|
Fourier series deterministic terms |
|
Constant and time trend determinstic terms based on calendar time |
|
Seasonal dummy deterministic terms based on calendar time |
|
Fourier series deterministic terms based on calendar time |
Abstract Base Class for all Deterministic Terms |
|
Abstract Base Class for calendar deterministic terms |
|
|
Abstract Base Class for all Fourier Deterministic Terms |
|
Abstract Base Class for all Time Trend Deterministic Terms |
Users who wish to write custom deterministic terms must subclass
DeterministicTerm
.
Abstract Base Class for all Deterministic Terms |
Forecasting Models¶
The Theta Model¶
The Theta model is a simple forecasting method that combines a linear time trend with a Simple Exponential Smoother (Assimakopoulos & Nikolopoulos). An estimator for the parameters of the Theta model and methods to forecast are available in:
|
The Theta forecasting model of Assimakopoulos and Nikolopoulos (2000) |
|
Results class from estimated Theta Models. |
Forecasting after STL Decomposition¶
statsmodels.tsa.seasonal.STL
is commonly used to remove seasonal
components from a time series. The deseasonalized time series can then
be modeled using a any non-seasonal model, and forecasts are constructed
by adding the forecast from the non-seasonal model to the estimates of
the seasonal component from the final full-cycle which are forecast using
a random-walk model.
|
Model-based forecasting using STL to remove seasonality |
|
Results for forecasting using STL to remove seasonality |
Prediction Results¶
Most forecasting methods support a get_prediction
method that return
a PredictionResults
object that contains both the prediction, its
variance and can construct a prediction interval.
Results Class¶
|
Prediction results |