WienerKhinchinTheorem - crowlogic/arb4j GitHub Wiki

Spectral decomposition theorem of stationary processes' autocorrelations

The Wiener–Khinchin theorem or Wiener–Khintchine theorem, also known as the Wiener–Khinchin–Einstein theorem or the Khinchin–Kolmogorov theorem, states that the autocorrelation function of a wide-sense-stationary random process has a spectral decomposition given by the power spectrum of that process.

The case of a continuous-time process

For continuous time, the Wiener–Khinchin theorem says that if $x$ is a wide-sense stochastic process whose autocorrelation function (sometimes called autocovariance) defined in terms of statistical expected value,

$$r_{xx}(\tau) = \mathbb{E}\big[x(t)^*x(t - \tau)\big]$$

exists and is finite at every lag $\tau$, then there exists a monotone function $F(f)$ in the frequency domain $-\infty < f < \infty$, or equivalently a non negative Radon measure $\mu$ on the frequency domain, such that

$$ r_{xx} (\tau) = \int_{-\infty}^\infty e^{2\pi i\tau f}\mu(df) = \int_{-\infty}^\infty e^{2\pi i\tau f} dF(f) , $$

where the integral is a Riemann–Stieltjes integral. The asterisk denotes complex conjugate, and can be omitted if the random process is real-valued. This is a kind of spectral decomposition of the auto-correlation function. $F$ is called the power spectral distribution function and is a genuine distribution function. It is also called the integrated spectrum.

The Fourier transform of $x(t)$ does not exist in general, because stochastic random functions are not generally neither square-integrable nor absolutely integrable. Nor is $r_{xx}$ assumed to be absolutely integrable, so it need not have a Fourier transform either.

However, if the measure $\mu(df) = dF(f)$ is absolutely continuous then $F$ is differentiable almost everywhere and we can write $\mu(df) = S(f) df$. In this case, one can determine $S(f)$, the power spectral density of $x(t)$, by taking the averaged derivative of $F$. Because the left and right derivatives of $F$ exist everywhere, i.e. we can write

$$S(f) = \frac12 \left(\lim_{\varepsilon \downarrow 0} \frac1\varepsilon \big(F(f + \varepsilon) - F(f)\big) + \lim_{\varepsilon \uparrow 0} \frac1\varepsilon \big(F(f + \varepsilon) - F(f)\big)\right)$$ everywhere, and the theorem simplifies to

$$ r_{xx} (\tau) = \int_{-\infty}^\infty e^{2\pi i\tau f} , S(f)df. $$

If now one assumes that r and S satisfy the necessary conditions for Fourier inversion to be valid, the Wiener–Khinchin theorem takes the simple form of saying that r and S are a Fourier-transform pair, and

$$ S(f) = \int_{-\infty}^\infty r_{xx} (\tau) e^{-2\pi if\tau} ,d\tau. $$

The case of a discrete-time process

For the discrete-time case, the power spectral density of the function with discrete values $x_n$ is

$$ S(\omega)=\frac{1}{2\pi} \sum_{k=-\infty}^\infty r_{xx}(k)e^{-i \omega k} $$

where $\omega = 2 \pi f$ is the angular frequency, $i$ is used to denote the imaginary unit (in engineering, sometimes the letter $j$ is used instead) and $r_{xx}(k)$ is the discrete autocorrelation function of $x_n$, defined in its deterministic or stochastic formulation.

Provided $r_{xx}$ is absolutely integrable, i.e.

$$ \sum_{k=-\infty}^\infty |r_{xx}(k)| < +\infty $$

the result of the theorem then writes

$$ r_{xx}(\tau) = \int_{-\pi}^{\pi} e^{i \tau \omega} S(\omega) d\omega $$

Being a sampled and discrete-time sequence, the spectral density is periodic in the frequency domain. This is due to the problem of aliasing: the contribution of any frequency higher than the Nyquist frequency seems to be equal to that of its alias between 0 and 1. For this reason, the domain of the function $S$ is usually restricted to $]-\pi, \pi]$ (note the interval is open from one side).

Application

The theorem is useful for analyzing linear time-invariant systems (LTI systems) when the inputs and outputs are not square-integrable, so their Fourier transforms do not exist. A corollary is that the Fourier transform of the autocorrelation function of the output of an LTI system is equal to the product of the Fourier transform of the autocorrelation function of the input of the system times the squared magnitude of the Fourier transform of the system impulse response.

Since the Fourier transform of the autocorrelation function of a signal is the power spectrum of the signal, this corollary is equivalent to saying that the power spectrum of the output is equal to the power spectrum of the input times the energy transfer function.

This corollary is used in the parametric method for power spectrum estimation.

Discrepancies in terminology

In many textbooks and in much of the technical literature it is tacitly assumed that Fourier inversion of the autocorrelation function and the power spectral density is valid, and the Wiener–Khinchin theorem is stated, very simply, as if it said that the Fourier transform of the autocorrelation function was equal to the power spectral density, ignoring all questions of convergence (Einstein is an example). But the theorem (as stated here) was applied by Norbert Wiener and Aleksandr Khinchin to the sample functions (signals) of wide-sense-stationary random processes, signals whose Fourier transforms do not exist. The whole point of Wiener's contribution was to make sense of the spectral decomposition of the autocorrelation function of a sample function of a wide-sense-stationary random process even when the integrals for the Fourier transform and Fourier inversion do not make sense.

Further complicating the issue is that the discrete Fourier transform always exists for digital, finite-length sequences, meaning that the theorem can be blindly applied to calculate auto-correlations of numerical sequences. As mentioned earlier, the relation of this discrete sampled data to a mathematical model is often misleading, and related errors can show up as a divergence when the sequence length is modified.

Some authors refer to $R$ as the autocovariance function. They then proceed to normalise it, by dividing by $R(0)$, to obtain what they refer to as the autocorrelation function.