The Gram–Charlier A series (named in honor of Jørgen Pedersen Gram and Carl Charlier), and the Edgeworth series (named in honor of Francis Ysidro Edgeworth) are series that approximate a probability distribution in terms of its cumulants.[1] The series are the same; but, the arrangement of terms (and thus the accuracy of truncating the series) differ.[2] The key idea of these expansions is to write the characteristic function of the distribution whose probability density function f is to be approximated in terms of the characteristic function of a distribution with known and suitable properties, and to recover f through the inverse Fourier transform.
We examine a continuous random variable. Let [math]\displaystyle{ \hat{f} }[/math] be the characteristic function of its distribution whose density function is f, and [math]\displaystyle{ \kappa_r }[/math] its cumulants. We expand in terms of a known distribution with probability density function ψ, characteristic function [math]\displaystyle{ \hat{\psi} }[/math], and cumulants [math]\displaystyle{ \gamma_r }[/math]. The density ψ is generally chosen to be that of the normal distribution, but other choices are possible as well. By the definition of the cumulants, we have (see Wallace, 1958)[3]
which gives the following formal identity:
By the properties of the Fourier transform, [math]\displaystyle{ (it)^r \hat{\psi}(t) }[/math] is the Fourier transform of [math]\displaystyle{ (-1)^r[D^r\psi](-x) }[/math], where D is the differential operator with respect to x. Thus, after changing [math]\displaystyle{ x }[/math] with [math]\displaystyle{ -x }[/math] on both sides of the equation, we find for f the formal expansion
If ψ is chosen as the normal density
with mean and variance as given by f, that is, mean [math]\displaystyle{ \mu = \kappa_1 }[/math] and variance [math]\displaystyle{ \sigma^2 = \kappa_2 }[/math], then the expansion becomes
since [math]\displaystyle{ \gamma_r=0 }[/math] for all r > 2, as higher cumulants of the normal distribution are 0. By expanding the exponential and collecting terms according to the order of the derivatives, we arrive at the Gram–Charlier A series. Such an expansion can be written compactly in terms of Bell polynomials as
Since the n-th derivative of the Gaussian function [math]\displaystyle{ \phi }[/math] is given in terms of Hermite polynomial as
this gives us the final expression of the Gram–Charlier A series as
Integrating the series gives us the cumulative distribution function
where [math]\displaystyle{ \Phi }[/math] is the CDF of the normal distribution.
If we include only the first two correction terms to the normal distribution, we obtain
with [math]\displaystyle{ He_3(x)=x^3-3x }[/math] and [math]\displaystyle{ He_4(x)=x^4 - 6x^2 + 3 }[/math].
Note that this expression is not guaranteed to be positive, and is therefore not a valid probability distribution. The Gram–Charlier A series diverges in many cases of interest—it converges only if [math]\displaystyle{ f(x) }[/math] falls off faster than [math]\displaystyle{ \exp(-(x^2)/4) }[/math] at infinity (Cramér 1957). When it does not converge, the series is also not a true asymptotic expansion, because it is not possible to estimate the error of the expansion. For this reason, the Edgeworth series (see next section) is generally preferred over the Gram–Charlier A series.
Edgeworth developed a similar expansion as an improvement to the central limit theorem.[4] The advantage of the Edgeworth series is that the error is controlled, so that it is a true asymptotic expansion.
Let [math]\displaystyle{ \{Z_i\} }[/math] be a sequence of independent and identically distributed random variables with finite mean [math]\displaystyle{ \mu }[/math] and variance [math]\displaystyle{ \sigma^2 }[/math], and let [math]\displaystyle{ X_n }[/math] be their standardized sums:
Let [math]\displaystyle{ F_n }[/math] denote the cumulative distribution functions of the variables [math]\displaystyle{ X_n }[/math]. Then by the central limit theorem,
for every [math]\displaystyle{ x }[/math], as long as the mean and variance are finite.
The standardization of [math]\displaystyle{ \{Z_i\} }[/math] ensures that the first two cumulants of [math]\displaystyle{ X_n }[/math] are [math]\displaystyle{ \kappa_1^{F_n} = 0 }[/math] and [math]\displaystyle{ \kappa_2^{F_n} = 1. }[/math] Now assume that, in addition to having mean [math]\displaystyle{ \mu }[/math] and variance [math]\displaystyle{ \sigma^2 }[/math], the i.i.d. random variables [math]\displaystyle{ Z_i }[/math] have higher cumulants [math]\displaystyle{ \kappa_r }[/math]. From the additivity and homogeneity properties of cumulants, the cumulants of [math]\displaystyle{ X_n }[/math] in terms of the cumulants of [math]\displaystyle{ Z_i }[/math] are for [math]\displaystyle{ r \geq 2 }[/math],
If we expand the formal expression of the characteristic function [math]\displaystyle{ \hat{f}_n(t) }[/math] of [math]\displaystyle{ F_n }[/math] in terms of the standard normal distribution, that is, if we set
then the cumulant differences in the expansion are
The Gram–Charlier A series for the density function of [math]\displaystyle{ X_n }[/math] is now
The Edgeworth series is developed similarly to the Gram–Charlier A series, only that now terms are collected according to powers of [math]\displaystyle{ n }[/math]. The coefficients of n−m/2 term can be obtained by collecting the monomials of the Bell polynomials corresponding to the integer partitions of m. Thus, we have the characteristic function as
where [math]\displaystyle{ P_j(x) }[/math] is a polynomial of degree [math]\displaystyle{ 3j }[/math]. Again, after inverse Fourier transform, the density function [math]\displaystyle{ f_n }[/math] follows as
Likewise, integrating the series, we obtain the distribution function
We can explicitly write the polynomial [math]\displaystyle{ P_m(-D) }[/math] as
where the summation is over all the integer partitions of m such that [math]\displaystyle{ \sum_i i k_i = m }[/math] and [math]\displaystyle{ l_i = i+2 }[/math] and [math]\displaystyle{ s = \sum_i k_i l_i. }[/math]
For example, if m = 3, then there are three ways to partition this number: 1 + 1 + 1 = 2 + 1 = 3. As such we need to examine three cases:
Thus, the required polynomial is
The first five terms of the expansion are[5]
Here, φ(j)(x) is the j-th derivative of φ(·) at point x. Remembering that the derivatives of the density of the normal distribution are related to the normal density by [math]\displaystyle{ \phi^{(n)}(x) = (-1)^n He_n(x)\phi(x) }[/math], (where [math]\displaystyle{ He_n }[/math] is the Hermite polynomial of order n), this explains the alternative representations in terms of the density function. Blinnikov and Moessner (1998) have given a simple algorithm to calculate higher-order terms of the expansion.
Note that in case of a lattice distributions (which have discrete values), the Edgeworth expansion must be adjusted to account for the discontinuous jumps between lattice points.[6]
Take [math]\displaystyle{ X_i \sim \chi^2(k=2), \, i=1, 2, 3 \, (n=3) }[/math] and the sample mean [math]\displaystyle{ \bar X = \frac{1}{3} \sum_{i=1}^{3} X_i }[/math].
We can use several distributions for [math]\displaystyle{ \bar X }[/math]:
Original source: https://en.wikipedia.org/wiki/Edgeworth series.
Read more |