The purely non-deterministic part is also called white noise . A linear deterministic component such as can be perfectly predicted based on its own past values (but can also contain random elements). The deterministic part can have a mean value that is constant over time, but also includes, for example, periodic, polynomial or exponential sequences in the points in time .
The required quadratic convergence of the series of guarantees the existence of the second moments of the process. For the validity of this decomposition no distribution assumptions have to be made and does not have to be independent; uncorrelatedness suffices .
The expected value is obtained
that is, the following applies:
The variance is calculated as follows:
Because of for this expression simplifies to
The variance is therefore finite and independent of time. According obtained with the autocovariances
with . It can be seen that the autocovariances are only a function of the time difference . Thus all conditions for the covariance stationarity are fulfilled. The autocorrelation function can be written as follows:
For example, ARMA models can be brought into Wold's representation. This representation is more of theoretical interest, because in practical applications models with an infinite number of parameters are useless.
Wold decomposition in functional analysis
There is also a Wold decomposition in functional analysis, see shift operator .
literature
Herman Wold A Study in the Analysis of Stationary Time Series , Stockholm: Almquist and Wicksell 1938
Gerhard Kirchgässner, Jürgen Wolters: Introduction to modern time series analysis , 1st edition, Munich: Vahlen, 2006, ISBN 978-3-800-63268-8 , p. 19f
↑ Used here in the sense that the covariance Cov ( ) only depends on the time difference (ts). Kirchgässner, Wolters, Hassner, Introduction to modern time series analysis, Springer 2013, p. 14 (weakly stationary is defined there as covariance and mean value stationary)