Derivation of the Partial Autocorrelation Function (PACF) for an AR(p) Process, demonstrating its Cut-off Property

Exploring the cinematic intuition of Derivation of the Partial Autocorrelation Function (PACF) for an AR(p) Process, demonstrating its Cut-off Property.

Visualizing...

Our institutional research engineers are currently mapping the formal proof for Derivation of the Partial Autocorrelation Function (PACF) for an AR(p) Process, demonstrating its Cut-off Property.

Apply for Institutional Early Access →

The Formal Theorem

Let {Xt} \{X_t\} be a stationary Autoregressive process of order p p , denoted AR(p), defined by the equation:
Xt=i=1pϕiXti+ϵt X_t = \sum_{i=1}^p \phi_i X_{t-i} + \epsilon_t
where ϕi \phi_i are the autoregressive coefficients with ϕp0 \phi_p \neq 0 , and ϵt \epsilon_t is a white noise process with E[ϵt]=0 E[\epsilon_t] = 0 and Var(ϵt)=σ2 \text{Var}(\epsilon_t) = \sigma^2 . The Partial Autocorrelation Function (PACF) at lag k k , denoted ϕkk \phi_{kk} , is the last coefficient of the best linear predictor of Xt X_t based on Xt1,,Xtk X_{t-1}, \dots, X_{t-k} . These coefficients, ϕk1,,ϕkk \phi_{k1}, \dots, \phi_{kk} , are derived by solving the Yule-Walker equations for an AR(k) model:
j=1kϕkjρlj=ρlfor l=1,,k \sum_{j=1}^k \phi_{kj} \rho_{|l-j|} = \rho_l \quad \text{for } l = 1, \dots, k
where ρl=Corr(Xt,Xtl) \rho_l = \text{Corr}(X_t, X_{t-l}) is the autocorrelation function (ACF) at lag l l . For an AR(p) process, the true ACF satisfies the Yule-Walker equations for l>p l > p :
ρl=i=1pϕiρlifor l>p \rho_l = \sum_{i=1}^p \phi_i \rho_{l-i} \quad \text{for } l > p
By substituting the vector (ϕ1,,ϕp,0,,0)T (\phi_1, \dots, \phi_p, 0, \dots, 0)^T (with kp k-p zeros) as a candidate solution into the Yule-Walker system for an AR(k) process where k>p k > p , it can be shown that these coefficients satisfy the system. This implies that for an AR(p) process, the PACF exhibits a sharp cut-off property:
ϕkk={0for kp0for k>p \phi_{kk} = \begin{cases} \neq 0 & \text{for } k \le p \\ 0 & \text{for } k > p \end{cases}
Specifically, ϕpp=ϕp \phi_{pp} = \phi_p , demonstrating that the direct linear dependence of Xt X_t on Xtk X_{t-k} , after accounting for intermediate observations, vanishes for lags beyond the true order p p of the AR process.

Analytical Intuition.

Imagine a grand symphony orchestra. Each instrument, Xti X_{t-i} , contributes to the current harmony, Xt X_t . In an AR(p) process, the conductor has specifically scored parts for only the most recent p p sections — say, the last p p measures of music. The 'partial autocorrelation' at lag k k , ϕkk \phi_{kk} , is like asking: 'How much unique influence does the instrument from k k measures ago, Xtk X_{t-k} , still have on the current note Xt X_t , *after* we've already listened to and accounted for all the instruments playing in the intervening k1 k-1 measures (Xt1,,Xtk+1 X_{t-1}, \dots, X_{t-k+1} )?' If our symphony truly only uses the last p p measures (an AR(p) process), then any instrument from k>p k > p measures ago, say Xt(p+1) X_{t-(p+1)} , would have its influence entirely 'filtered out' or 'explained away' by the p p crucial preceding instruments. Its unique, direct contribution becomes zero, like a silent instrument beyond its assigned score. This vanishing act is the PACF's cinematic cut-off property.
CAUTION

Institutional Warning.

Students often confuse PACF with ACF. While ACF measures total correlation, PACF isolates the *direct* effect, net of intermediate influences. Misinterpreting sample PACF values beyond p p as small but significant rather than statistically zero is a common pitfall.

Academic Inquiries.

01

What is the practical significance of the PACF cut-off property?

The cut-off property is crucial for identifying the order p p of an AR(p) model. By inspecting the sample PACF plot, one looks for the lag k k after which the PACF values become statistically insignificant (close to zero). This lag k k indicates the likely order p p of the AR component.

02

How does the PACF differ from the ACF for an AR(p) process?

For an AR(p) process, the ACF ρk \rho_k typically decays exponentially or sinusoidally (tails off) as k k increases, meaning all past observations have some indirect influence. In contrast, the PACF ϕkk \phi_{kk} *cuts off* abruptly to zero after lag p p , indicating that only the most recent p p observations have a *direct* impact on the current value.

03

Is the cut-off property always perfectly observed in real-world data?

In theoretical AR(p) processes, the cut-off is exact. However, with real-world time series data, we work with *sample* PACF values. These will rarely be exactly zero due to sampling variability. We typically look for PACF values that fall within a confidence interval around zero for lags k>p k > p to declare a cut-off, not necessarily exact zeros.

04

Can PACF also be used for MA(q) processes?

Yes, for an MA(q) process, the PACF theoretically decays exponentially or sinusoidally (tails off), while the ACF exhibits a cut-off at lag q q . This complementary behavior of ACF and PACF helps distinguish between AR and MA components when identifying suitable ARIMA models.

Standardized References.

  • Definitive Institutional SourceBrockwell, Peter J., and Richard A. Davis. Introduction to Time Series and Forecasting.

Institutional Citation

Reference this proof in your academic research or publications.

NICEFA Visual Mathematics. (2026). Derivation of the Partial Autocorrelation Function (PACF) for an AR(p) Process, demonstrating its Cut-off Property: Visual Proof & Intuition. Retrieved from https://nicefa.org/library/time-series-analysis/derivation-of-the-partial-autocorrelation-function--pacf--for-an-ar-p--process--demonstrating-its-cut-off-property

Dominate the Logic.

"Abstract theory is just a movement we haven't seen yet."