Martingale difference sequence

In probability theory, a martingale difference sequence (MDS) is related to the concept of the martingale. A stochastic series X is an MDS if its expectation with respect to the past is zero. Formally, consider an adapted sequence $$\{X_t, \mathcal{F}_t\}_{-\infty}^{\infty}$$ on a probability space $$(\Omega, \mathcal{F}, \mathbb{P})$$. $$X_t$$ is an MDS if it satisfies the following two conditions:


 * $$ \mathbb{E} \left|X_t\right| < \infty $$, and


 * $$ \mathbb{E} \left[X_t | \mathcal{F}_{t-1}\right] = 0, a.s. $$,

for all $$t$$. By construction, this implies that if $$Y_t$$ is a martingale, then $$X_t=Y_t-Y_{t-1}$$ will be an MDS—hence the name.

The MDS is an extremely useful construct in modern probability theory because it implies much milder restrictions on the memory of the sequence than independence, yet most limit theorems that hold for an independent sequence will also hold for an MDS.

A special case of MDS, denoted as {Xt,$$ \mathcal{F}$$t}0${\infty} $ is known as innovative sequence of Sn; where Sn and $$ \mathcal{F}_{t}$$ are corresponding to random walk and filtration of the random processes $$\{X_{t}\}_0^\infty $$.

In probability theory innovation series is used to emphasize the generality of Doob representation. In signal processing the innovation series is used to introduce Kalman filter. The main differences of innovation terminologies are in the applications. The later application aims to introduce the nuance of samples to the model by random sampling.