Isserlis' theorem rewrites higher-order moments of zero-mean Gaussian random variables as a sum of nd-order moments of pairings. For zero-mean Gaussian random variables , Isserlis' theorem tells us for even
where denotes a pair of indices in . There are terms in the sum.
For odd , . Note that the nd-order moments are also covariances for zero-mean random variables: .
As an example, Isserlis' theorem applied to the -th order moments goes like this
Proof from Scratch
We have the random variable , which follows a zero-mean Gaussian distribution .
We use a quadratic identity equation
Because the probability density function integrate to , we have
Then we make the genius move of differentiating Equation with respect to and evaluating at . Differentiating the right-hand side, we get exactly the moments we want
Differentiating the left-hand side, we get
This is getting really cumbersome to write. I am lazy; so let's just assume as an example of the even case. We continue the derivation for
Note that only the terms with in the sum have nonzero derivatives with respect to . There are arrangements for . There are permutations for , namely . Each permutation of appears times. We thus write out Equation as
By the definition of the covariance matrix, we finish the proof
Although I only completed the derivation for , we can generalize the result to arbitrary even by induction. For instance, for , we can first break things down to and and then break down the . Namely, we first have
The -th order moment in each term can be further decomposed to three terms that only contain nd moments using Equation . So we eventually break the -th order moment into terms.
Proof via Stein's Lemma
If one already knows Stein's Lemma, Isserlis' theorem is quite easy to prove. Stein's lemma states that for zero-mean Gaussian random variables , we have
Applying this lemma to immediately gives us Equation
Comments
Isserlis' theorem is specific to zero-mean Gaussian random variables. It does not extend to other distributions.
Isserlis' theorem allows the random variables to have correlations. If the random variables are independent , we have no matter what distributions and follow.