Hubbry Logo
Autoregressive moving-average modelAutoregressive moving-average modelMain
Open search
Autoregressive moving-average model
Community hub
Autoregressive moving-average model
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Autoregressive moving-average model
Autoregressive moving-average model
from Wikipedia

In the statistical analysis of time series, an autoregressive–moving-average (ARMA) model is used to represent a (weakly) stationary stochastic process by combining two components: autoregression (AR) and moving average (MA). These models are widely used for analyzing the structure of a series and for forecasting future values.

The AR component specifies that the current value of the series depends linearly on its own past values (lags), while the MA component specifies that the current value depends on a linear combination of past error terms. An ARMA model is typically denoted as ARMA(p, q), where p is the order of the autoregressive part and q is the order of the moving-average part.

The general ARMA model was described in the 1951 thesis of Peter Whittle, Hypothesis testing in time series analysis, and it was popularized in the 1970 book by George E. P. Box and Gwilym Jenkins.

ARMA models can be estimated by using the Box–Jenkins method.

Mathematical formulation

[edit]

Autoregressive model

[edit]

The notation AR(p) refers to the autoregressive model of order p. The AR(p) model is written as

where are parameters and the random variable is white noise, usually independent and identically distributed (i.i.d.) normal random variables.[1][2]

In order for the model to remain stationary, the roots of its characteristic polynomial must lie outside the unit circle. For example, processes in the AR(1) model with are not stationary because the root of lies within the unit circle.[3]

The augmented Dickey–Fuller test can assesses the stability of an intrinsic mode function and trend components. For stationary time series, the ARMA models can be used, while for non-stationary series, Long short-term memory models can be used to derive abstract features. The final value is obtained by reconstructing the predicted outcomes of each time series.[citation needed]

Moving average model

[edit]

The notation MA(q) refers to the moving average model of order q:

where the are the parameters of the model, is the expectation of (often assumed to equal 0), and , ..., are i.i.d. white noise error terms that are commonly normal random variables.[4]

ARMA model

[edit]

The notation ARMA(p, q) refers to the model with p autoregressive terms and q moving-average terms. This model contains the AR(p) and MA(q) models,[5]

In terms of lag operator

[edit]

In some texts, the models is specified using the lag operator L. In these terms, the AR(p) model is given by

where represents the polynomial

The MA(q) model is given by

where represents the polynomial

Finally, the combined ARMA(p, q) model is given by

or more concisely,

or

This is the form used in Box, Jenkins & Reinsel.[6]

Moreover, starting summations from and setting and , then we get an even more elegant formulation:

Spectrum

[edit]

The spectral density of an ARMA process iswhere is the variance of the white noise, is the characteristic polynomial of the moving average part of the ARMA model, and is the characteristic polynomial of the autoregressive part of the ARMA model.[7][8]

Fitting models

[edit]

Choosing p and q

[edit]

An appropriate value of p in the ARMA(p, q) model can be found by plotting the partial autocorrelation functions. Similarly, q can be estimated by using the autocorrelation functions. Both p and q can be determined simultaneously using extended autocorrelation functions (EACF).[9] Further information can be gleaned by considering the same functions for the residuals of a model fitted with an initial selection of p and q.

Brockwell & Davis recommend using Akaike information criterion (AIC) for finding p and q.[10] Another option is the Bayesian information criterion (BIC).

Estimating coefficients

[edit]

After choosing p and q, ARMA models can be fitted by least squares regression to find the values of the parameters which minimize the error term. It is good practice to find the smallest values of p and q which provide an acceptable fit to the data. For a pure AR model, the Yule-Walker equations may be used to provide a fit.

ARMA outputs are used primarily to forecast (predict), and not to infer causation as in other areas of econometrics and regression methods such as OLS and 2SLS.

Software implementations

[edit]
  • In R, standard package stats has function arima, documented in ARIMA Modelling of Time Series. Package astsa has an improved script called sarima for fitting ARMA models (seasonal and nonseasonal) and sarima.sim to simulate data from these models. Extension packages contain related and extended functionality: package tseries includes the function arma(), documented in "Fit ARMA Models to Time Series"; packagefracdiff contains fracdiff() for fractionally integrated ARMA processes; and package forecast includes auto.arima for selecting a parsimonious set of p, q. The CRAN task view on Time Series contains links to most of these.
  • Mathematica has a complete library of time series functions including ARMA.[11]
  • MATLAB includes functions such as arma, ar and arx to estimate autoregressive, exogenous autoregressive and ARMAX models. See System Identification Toolbox and Econometrics Toolbox for details.
  • Julia has community-driven packages that implement fitting with an ARMA model such as arma.jl.
  • Python has the statsmodelsS package which includes many models and functions for time series analysis, including ARMA. Formerly part of the scikit-learn library, it is now stand-alone and integrates well with Pandas.
  • PyFlux has a Python-based implementation of ARIMAX models, including Bayesian ARIMAX models.
  • IMSL Numerical Libraries are libraries of numerical analysis functionality including ARMA and ARIMA procedures implemented in standard programming languages like C, Java, C# .NET, and Fortran.
  • gretl can estimate ARMA models, as mentioned here
  • GNU Octave extra package octave-forge supports AR models.
  • Stata includes the function arima. for ARMA and ARIMA models.
  • SuanShu is a Java library of numerical methods that implements univariate/multivariate ARMA, ARIMA, ARMAX, etc models, documented in "SuanShu, a Java numerical and statistical library".
  • SAS has an econometric package, ETS, that estimates ARIMA models. See details.

History and interpretations

[edit]

The general ARMA model was described in the 1951 thesis of Peter Whittle, who used mathematical analysis (Laurent series and Fourier analysis) and statistical inference.[12][13] ARMA models were popularized by a 1970 book by George E. P. Box and Jenkins, who expounded an iterative (Box–Jenkins) method for choosing and estimating them. This method was useful for low-order polynomials (of degree three or less).[14]

ARMA is essentially an infinite impulse response filter applied to white noise, with some additional interpretation placed on it.

In digital signal processing, ARMA is represented as a digital filter with white noise at the input and the ARMA process at the output.

Applications

[edit]

ARMA is appropriate when a system is a function of a series of unobserved shocks (the MA or moving average part) as well as its own behavior. For example, stock prices may be shocked by fundamental information as well as exhibiting technical trending and mean-reversion effects due to market participants.[citation needed]

Generalizations

[edit]

There are various generalizations of ARMA. Nonlinear AR (NAR), nonlinear MA (NMA) and nonlinear ARMA (NARMA) model nonlinear dependence on past values and error terms. Vector AR (VAR) and vector ARMA (VARMA) model multivariate time series. Autoregressive integrated moving average (ARIMA) models non-stationary time series (that is, whose mean changes over time). Autoregressive conditional heteroskedasticity (ARCH) models time series where the variance changes. Seasonal ARIMA (SARIMA or periodic ARMA) models periodic variation. Autoregressive fractionally integrated moving average (ARFIMA, or Fractional ARIMA, FARIMA) model time-series that exhibits long memory. Multiscale AR (MAR) is indexed by the nodes of a tree instead of integers.

Autoregressive–moving-average model with exogenous inputs (ARMAX)

[edit]

The notation ARMAX(p, q, b) refers to a model with p autoregressive terms, q moving average terms and b exogenous inputs terms. The last term is a linear combination of the last b terms of a known and external time series . It is given by:

where are the parameters of the exogenous input .

Some nonlinear variants of models with exogenous variables have been defined: see for example Nonlinear autoregressive exogenous model.

Statistical packages implement the ARMAX model through the use of "exogenous" (that is, independent) variables. Care must be taken when interpreting the output of those packages, because the estimated parameters usually (for example, in R[15] and gretl) refer to the regression:

where incorporates all exogenous (or independent) variables:

See also

[edit]

References

[edit]

Further reading

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
The autoregressive moving-average (ARMA) model is a statistical framework used in time series analysis to represent stationary processes by combining autoregressive (AR) and moving average (MA) components, enabling the modeling of linear dependencies in data over time. In an ARMA(p, q) model, the AR part of order p expresses the current observation as a of its p previous values, while the MA part of order q incorporates the influence of the q preceding forecast errors, with the process driven by innovations. The model's general equation is: xt=i=1pϕixti+ϵt+j=1qθjϵtj,x_t = \sum_{i=1}^p \phi_i x_{t-i} + \epsilon_t + \sum_{j=1}^q \theta_j \epsilon_{t-j}, where xtx_t is the time series value at time t, ϕi\phi_i are the AR coefficients, θj\theta_j are the MA coefficients, and {ϵt}\{\epsilon_t\} is a sequence of independent and identically distributed random errors with mean zero and constant variance (white noise). This formulation assumes weak stationarity, meaning the mean, variance, and autocovariance structure of the series remain constant over time, which requires the roots of the AR and MA characteristic polynomials to lie outside the unit circle. ARMA models were formalized and popularized through the seminal work of statisticians and Gwilym M. Jenkins in their 1970 book Time Series Analysis: Forecasting and Control, which outlined an iterative methodology for model building involving identification via and partial autocorrelation functions, parameter (often via maximum likelihood), and diagnostic checking using residual . This Box-Jenkins approach revolutionized by emphasizing data-driven over ad hoc methods. Key properties include invertibility for the MA component, allowing representation as an infinite AR process, and the ability to capture both short-term persistence (via AR) and smoothing effects (via MA) in univariate stationary series. In practice, ARMA models underpin applications in for forecasting GDP or , for stock returns analysis, and engineering for , though they are typically extended to models when data exhibit trends or non-stationarity through differencing. Recent advancements include variants like generalized ARMA for non-Gaussian errors and space-time extensions for spatiotemporal data, enhancing their utility in modern contexts such as and integrations.

Model Components

Autoregressive Process

The autoregressive (AR) process of order pp, denoted AR(pp), is a fundamental model in time series analysis that describes how the current value of a series depends linearly on its own previous pp values, plus a random shock. This structure captures the endogenous dependence within the series, making it suitable for modeling phenomena with inertia or momentum, such as economic indicators or financial returns. The mathematical formulation of an AR(pp) process is given by yt=c+i=1pϕiyti+ϵt,y_t = c + \sum_{i=1}^p \phi_i y_{t-i} + \epsilon_t, where cc represents a (often related to the of the process), the coefficients ϕi\phi_i (for i=1,,pi = 1, \dots, p) quantify the influence of each lagged value, and ϵt\epsilon_t is a error term with zero , constant variance σ2>0\sigma^2 > 0, and no serial correlation (i.e., E[ϵtϵs]=0\mathbb{E}[\epsilon_t \epsilon_s] = 0 for tst \neq s). The parameters ϕi\phi_i must satisfy specific conditions to ensure the process behaves in a manner over time. For the AR(pp) process to be stationary—meaning its statistical properties like mean and variance remain constant over time—the roots of the characteristic polynomial Φ(z)=1i=1pϕizi=0\Phi(z) = 1 - \sum_{i=1}^p \phi_i z^i = 0 must lie outside the unit circle in the complex plane, i.e., all roots zkz_k satisfy zk>1|z_k| > 1. This condition ensures that the effects of past shocks do not accumulate indefinitely, preventing explosive behavior or non-constant variance. If any root has modulus less than or equal to 1, the process becomes non-stationary, often exhibiting trends or unit root behavior. A simple yet illustrative example is the AR(1) process, yt=c+ϕyt1+ϵty_t = c + \phi y_{t-1} + \epsilon_t, where stationarity requires ϕ<1|\phi| < 1. Here, ϕ\phi directly measures the degree of persistence: if ϕ\phi is close to 1 (e.g., 0.9), shocks to the series decay slowly, leading to prolonged deviations from the mean and high autocorrelation; conversely, if ϕ\phi is near 0, the series behaves more like white noise with minimal memory. This persistence is crucial for understanding phenomena like business cycles, where ϕ0.8\phi \approx 0.8–0.95 is common in empirical macroeconomic data. Under the stationarity condition, a causal AR(pp) process admits an infinite moving average (MA) representation, expressing yty_t as an infinite linear combination of current and past white noise terms. This Wold decomposition is derived by inverting the AR operator: starting from the AR equation, recursive substitution yields yt=μ+j=0ψjϵtjy_t = \mu + \sum_{j=0}^\infty \psi_j \epsilon_{t-j}, where μ=c/(1i=1pϕi)\mu = c / (1 - \sum_{i=1}^p \phi_i) is the process mean, and the ψj\psi_j coefficients are determined by the expansion of (1i=1pϕiLi)1(1 - \sum_{i=1}^p \phi_i L^i)^{-1} (with LL the lag operator), satisfying j=0ψj<\sum_{j=0}^\infty |\psi_j| < \infty for convergence. For the AR(1) case, the derivation is straightforward: ytϕyt1=c+ϵty_t - \phi y_{t-1} = c + \epsilon_t, iterating backward gives yt=ck=0ϕk+j=0ϕjϵtjy_t = c \sum_{k=0}^\infty \phi^k + \sum_{j=0}^\infty \phi^j \epsilon_{t-j}, with ψj=ϕj\psi_j = \phi^j and μ=c/(1ϕ)\mu = c / (1 - \phi), illustrating how past shocks propagate with geometrically decaying weights. This representation underscores the AR process's equivalence to an infinite-order MA under stationarity, facilitating forecasting and spectral analysis.

Moving Average Process

The moving average process of order qq, denoted MA(qq), models a time series where each observation is a constant plus the current white noise error term and a finite linear combination of the previous qq error terms. This structure captures how past forecast errors influence current values, reflecting temporary shocks with limited persistence. The mathematical formulation is yt=μ+ϵt+i=1qθiϵti,y_t = \mu + \epsilon_t + \sum_{i=1}^q \theta_i \epsilon_{t-i}, where μ\mu is the process mean (often zero for centered series), the θi\theta_i (for i=1,,qi = 1, \dots, q) are fixed moving average coefficients that weight the impact of past errors, and {ϵt}\{ \epsilon_t \} is white noise—a sequence of i.i.d. random variables with E(ϵt)=0E(\epsilon_t) = 0 and Var(ϵt)=σ2>0\text{Var}(\epsilon_t) = \sigma^2 > 0, typically assumed Gaussian for exact . The parameters θi\theta_i and σ2\sigma^2 are estimated from , and the model assumes no further dependence beyond the specified lags. By construction, the MA(qq) process is always (weakly) stationary, as its mean is constant and autocovariances depend only on the lag, owing to the finite summation of stationary white noise components. However, invertibility—a condition ensuring the process can be expressed as an infinite autoregressive form for practical forecasting—requires that all roots of the MA polynomial θ(z)=1+i=1qθizi=0\theta(z) = 1 + \sum_{i=1}^q \theta_i z^i = 0 lie outside the unit circle (z>1|z| > 1) in the complex plane. Non-invertible models, while mathematically valid, complicate estimation and interpretation, so invertible parameterizations are preferred. A basic illustration is the MA(1) model, yt=μ+ϵt+θ1ϵt1y_t = \mu + \epsilon_t + \theta_1 \epsilon_{t-1}, which models dependence limited to the immediate past error. This form is particularly effective for short-term dynamics, such as fleeting market shocks in , where only the most recent error meaningfully affects the current observation. When θ11\theta_1 \approx -1, the model can approximate over-differencing effects, where excessive differencing of a stationary series induces artificial negative lag-1 that the MA term corrects. The autocorrelation function (ACF) of an MA(qq) process is derived from its autocovariance structure and truncates exactly after lag qq, a hallmark for model identification. The lag-kk autocovariance is γk=E[(ytμ)(ytkμ)]=σ2j=0qkθjθj+k\gamma_k = E[(y_t - \mu)(y_{t-k} - \mu)] = \sigma^2 \sum_{j=0}^{q-k} \theta_j \theta_{j+k} for k=1,,qk = 1, \dots, q (with θ0=1\theta_0 = 1 and θj=0\theta_j = 0 for j>qj > q), while γ0=σ2(1+i=1qθi2)\gamma_0 = \sigma^2 (1 + \sum_{i=1}^q \theta_i^2) and γk=0\gamma_k = 0 for k>qk > q. The autocorrelations follow as ρk=γk/γ0\rho_k = \gamma_k / \gamma_0, yielding nonzero values only up to lag qq, which reflects the process's finite memory. This cutoff pattern contrasts with processes having longer-range dependence. The MA process complements autoregressive models by focusing on error-driven correlations rather than value persistence.

ARMA Formulation

General ARMA Equation

The autoregressive moving-average (ARMA) model of order (p, q), denoted ARMA(p, q), provides a unified framework for modeling stationary time series by integrating autoregressive and moving average components, allowing representation of processes with both lagged dependencies in observations and errors. This approach builds briefly on the pure autoregressive and moving average processes as foundational building blocks. The general equation for an ARMA(p, q) process with mean μ\mu is (ytμ)i=1pϕi(ytiμ)=ϵt+i=1qθiϵti,(y_t - \mu) - \sum_{i=1}^p \phi_i (y_{t-i} - \mu) = \epsilon_t + \sum_{i=1}^q \theta_i \epsilon_{t-i}, where yty_t is the value at time tt, {ϵt}\{\epsilon_t\} is a sequence of errors with zero and constant variance σ2\sigma^2, the coefficients ϕ1,,ϕp\phi_1, \dots, \phi_p are the autoregressive parameters, and θ1,,θq\theta_1, \dots, \theta_q are the parameters. For centered processes where μ=0\mu = 0, the equation simplifies to yti=1pϕiyti=ϵt+i=1qθiϵti.y_t - \sum_{i=1}^p \phi_i y_{t-i} = \epsilon_t + \sum_{i=1}^q \theta_i \epsilon_{t-i}. Here, pp represents the degree of the autoregressive component, indicating the number of prior observations influencing the current value, while qq represents the degree of the moving average component, indicating the number of prior errors affecting the current value. Validity of the model requires stationarity of the AR part, ensured by roots of the associated characteristic polynomial lying outside the unit circle, and invertibility of the MA part, similarly requiring roots outside the unit circle to allow expression as an infinite AR process. A compact notation using the backshift (lag) operator BB, defined such that Byt=yt1B y_t = y_{t-1}, expresses the model as ϕ(B)(ytμ)=θ(B)ϵt\phi(B)(y_t - \mu) = \theta(B) \epsilon_t, where ϕ(B)=1i=1pϕiBi\phi(B) = 1 - \sum_{i=1}^p \phi_i B^i and θ(B)=1+i=1qθiBi\theta(B) = 1 + \sum_{i=1}^q \theta_i B^i, though full analysis of this form follows separately. Special cases of the ARMA(p, q) model include the pure autoregressive AR(p) when q=0q = 0, reducing to yti=1pϕiyti=ϵty_t - \sum_{i=1}^p \phi_i y_{t-i} = \epsilon_t (assuming μ=0\mu = 0), which emphasizes dependence solely on past values. The pure moving average MA(q) arises when p=0p = 0, given by yt=ϵt+i=1qθiϵtiy_t = \epsilon_t + \sum_{i=1}^q \theta_i \epsilon_{t-i}, focusing on finite dependence on past shocks. A representative example is the ARMA(1,1) model, ytϕ1yt1=ϵt+θ1ϵt1y_t - \phi_1 y_{t-1} = \epsilon_t + \theta_1 \epsilon_{t-1}, which models series exhibiting persistence from the prior observation tempered by a single lagged error, commonly applied to capture exponential decay in autocorrelations.

Lag Operator Representation

The , often denoted LL and also known as the backshift operator, provides a compact notation for expressing time shifts in time series models, defined such that Lyt=yt1L y_t = y_{t-1}. Higher powers extend this action linearly, with Lkyt=ytkL^k y_t = y_{t-k} for any positive kk. This operator facilitates the representation of linear combinations through , such as the autoregressive ϕ(L)=1i=1pϕiLi\phi(L) = 1 - \sum_{i=1}^p \phi_i L^i and the θ(L)=1+i=1qθiLi\theta(L) = 1 + \sum_{i=1}^q \theta_i L^i, where the coefficients ϕi\phi_i and θi\theta_i characterize the dependencies in the series. In this framework, the general autoregressive (ARMA) model of orders pp and qq is succinctly written as ϕ(L)yt=θ(L)ϵt\phi(L) y_t = \theta(L) \epsilon_t, where ϵt\epsilon_t is with mean zero and variance σ2\sigma^2. This operator form highlights the multiplicative structure of the polynomials, allowing for elegant manipulations in theoretical derivations. For a stationary ARMA process—requiring the roots of ϕ(z)=0\phi(z) = 0 to lie outside the unit circle—the model admits an infinite (MA) representation: yt=j=0ψjϵtjy_t = \sum_{j=0}^\infty \psi_j \epsilon_{t-j}, where the weights ψj\psi_j are generated by the power of ψ(L)=θ(L)/ϕ(L)\psi(L) = \theta(L) / \phi(L), with ψ0=1\psi_0 = 1. An invertible ARMA process similarly yields an infinite autoregressive (AR) form. The lag operator notation offers significant advantages in time series analysis, including streamlined simulation of processes by recursively applying the operator to generate future values, efficient forecasting through recursive computation of expectations (e.g., E[yt+hFt]=j=hψjϵt+hjE[y_{t+h} | \mathcal{F}_t] = \sum_{j=h}^\infty \psi_j \epsilon_{t+h-j} for the MA(\infty) representation), and derivation of key moments such as the autocovariance function via polynomial expansions. It also simplifies proofs of properties like ergodicity under stationarity assumptions. For illustration, consider a autoregressive AR(1) model, ytϕ1yt1=ϵty_t - \phi_1 y_{t-1} = \epsilon_t, which in lag notation becomes (1ϕ1L)yt=ϵt(1 - \phi_1 L) y_t = \epsilon_t with ϕ1<1|\phi_1| < 1 for stationarity. Similarly, a moving average MA(1) model, yt=ϵt+θ1ϵt1y_t = \epsilon_t + \theta_1 \epsilon_{t-1}, is expressed as yt=(1+θ1L)ϵty_t = (1 + \theta_1 L) \epsilon_t, with invertibility requiring θ1<1|\theta_1| < 1. These forms reveal the AR(1) as an MA(\infty) process, yt=j=0ϕ1jϵtjy_t = \sum_{j=0}^\infty \phi_1^j \epsilon_{t-j}.

Model Properties

Stationarity and Invertibility

In time series analysis, strict stationarity refers to a stochastic process where the joint distribution of any collection of observations is invariant to time shifts, implying constant mean, variance, and autocovariances that depend solely on the lag between observations. For ARMA models, the focus is typically on weak (or second-order) stationarity, which requires a time-invariant mean and autocovariance function, ensuring the process has finite second moments and is suitable for modeling with constant parameters. For an ARMA(p, q) process defined by the equation ϕ(B)yt=θ(B)ϵt\phi(B) y_t = \theta(B) \epsilon_t, where ϕ(z)=1ϕ1zϕpzp\phi(z) = 1 - \phi_1 z - \cdots - \phi_p z^p and θ(z)=1+θ1z++θqzq\theta(z) = 1 + \theta_1 z + \cdots + \theta_q z^q are the autoregressive (AR) and moving average (MA) polynomials, respectively, stationarity holds if all roots of the characteristic equation ϕ(z)=0\phi(z) = 0 lie strictly outside the unit circle in the complex plane (i.e., have absolute values greater than 1). These roots determine the behavior of the process: when they are outside the unit circle, the AR component exhibits mean reversion, as past shocks decay over time, leading to a stable process with bounded variance. Conversely, if any root lies inside the unit circle (absolute value less than 1), the process becomes explosive, with variance growing without bound and forecasts diverging rapidly. Roots on the unit circle (absolute value equal to 1), such as a unit root in an AR(1) model where ϕ1=1\phi_1 = 1, result in non-stationarity, manifesting as persistent trends or random walk-like behavior without mean reversion. Invertibility, a complementary property, ensures the MA component can be expressed as an infinite AR process, facilitating practical forecasting and parameter estimation. It requires all roots of θ(z)=0\theta(z) = 0 to lie outside the unit circle, mirroring the stationarity condition but applied to the MA polynomial. This condition guarantees that current observations depend on past errors in a decaying manner, avoiding non-unique representations of the process. Non-stationarity in ARMA models, particularly due to unit roots, violates the constant mean and variance assumptions, leading to unreliable parameter estimates, spurious regressions, and forecasts that fail to capture long-term dynamics. In such cases, transformations like differencing are necessary, extending the model to ARIMA frameworks to induce stationarity. Testing for stationarity conceptually involves examining the roots of the AR polynomial or applying unit root tests, which assess the null hypothesis of a unit root against the alternative of stationarity, often through augmented regressions to account for serial correlation.

Autocorrelation Structure

The autocorrelation function (ACF) and partial autocorrelation function (PACF) characterize the serial correlation structure of stationary ARMA processes, providing key patterns for model identification. For a stationary ARMA(p, q) process, the ACF measures the correlation between observations separated by lag k, while the PACF isolates the correlation at lag k after adjusting for intermediate lags. In a pure autoregressive AR(p) process, the ACF decays exponentially or in a damped sinusoidal manner as the lag increases, reflecting persistent dependence on past values. The autocorrelations satisfy the Yule-Walker equations: for k > 0, ρk=i=1pϕiρki\rho_k = \sum_{i=1}^p \phi_i \rho_{k-i}, where ρk\rho_k is the at lag k and ϕi\phi_i are the AR coefficients. In contrast, the PACF for AR(p) truncates to zero after lag p, showing no significant beyond the order. For a pure moving average MA(q) process, the ACF truncates abruptly to zero after lag q, as correlations depend only on the finite shock history. The PACF, however, decays gradually without truncation, similar to the ACF of an AR process. In a mixed ARMA(p, q) process, the ACF and PACF exhibit hybrid behaviors: the ACF typically shows a non-zero pattern up to lag q followed by exponential decay influenced by the AR component, while the PACF decays without clear truncation. These mixed patterns distinguish ARMA models from pure AR or MA, aiding in order selection during identification. For example, in an AR(1) model yt=ϕ1yt1+ϵty_t = \phi_1 y_{t-1} + \epsilon_t with ϕ1<1|\phi_1| < 1, the ACF plot displays ρk=ϕ1k\rho_k = \phi_1^k, a smooth exponential decay from lag 1 onward, while the PACF spikes at lag 1 (ϕ11=ϕ1\phi_{11} = \phi_1) and drops to zero thereafter. In an MA(1) model yt=ϵt+θ1ϵt1y_t = \epsilon_t + \theta_1 \epsilon_{t-1} with θ1<1|\theta_1| < 1, the ACF has ρ1=θ1/(1+θ12)\rho_1 = \theta_1 / (1 + \theta_1^2) at lag 1 and zero beyond, whereas the PACF decays exponentially. For an ARMA(1,1) model yt=ϕ1yt1+ϵt+θ1ϵt1y_t = \phi_1 y_{t-1} + \epsilon_t + \theta_1 \epsilon_{t-1}, the ACF features a distinct ρ1=(ϕ1+θ1)(1+ϕ1θ1)1+θ12+2ϕ1θ1\rho_1 = \frac{(\phi_1 + \theta_1)(1 + \phi_1 \theta_1)}{1 + \theta_1^2 + 2 \phi_1 \theta_1} followed by geometric decay ρk=ϕ1ρk1\rho_k = \phi_1 \rho_{k-1} for k ≥ 2, and the PACF decays gradually, blending the truncation and persistence of its components.

Spectral Density

The power spectral density (PSD) of a weakly stationary time series process is defined as the Fourier transform of its autocovariance function, providing a frequency-domain representation of the process's variance distribution across frequencies. For an ARMA(p, q) process defined by ϕ(B)Yt=θ(B)ϵt\phi(B) Y_t = \theta(B) \epsilon_t, where {ϵt}\{\epsilon_t\} is white noise with variance σ2\sigma^2, ϕ(z)=1j=1pϕjzj\phi(z) = 1 - \sum_{j=1}^p \phi_j z^j, and θ(z)=1+j=1qθjzj\theta(z) = 1 + \sum_{j=1}^q \theta_j z^j, the PSD is given by f(ω)=σ22πθ(eiω)ϕ(eiω)2,ω[π,π].f(\omega) = \frac{\sigma^2}{2\pi} \left| \frac{\theta(e^{-i\omega})}{\phi(e^{-i\omega})} \right|^2, \quad \omega \in [-\pi, \pi].
Add your contribution
Related Hubs
User Avatar
No comments yet.