Professional Documents
Culture Documents
A BSTRACT. In this paper we introduce new asymptotic expansions for probability func-
tions of sums of independent and identically distributed random variables. Results are
obtained by efficiently employing information provided by lower-order convolutions.
In comparison with Edgeworth-type theorems, advantages include improved asymp-
totic results in the case of symmetric random variables and ease of computation of main
error terms and asymptotic crossing points. The first-order estimate can perform quite
well against the corresponding renormalized saddlepoint approximation and, pointwise,
requires evaluation of only a single convolution integral. While the new expansions are
fairly straightforward, the implications are fortuitous and may spur further related work.
1. I NTRODUCTION
In this paper we introduce new asymptotic expansions for probability functions of
sums of independent and identically distributed (i.i.d.) random variables, employing in-
formation provided by lower-order convolutions. In particular, suppose X is a random
variable with mean and finite variance 2 , and X1 , X2 , . . . is a sequence of i.i.d. ran-
dom variables each with the same distribution as X. In addition, suppose that X has
probability function f , where f is either a density or a probability mass function. Let
f (i) be the probability function for the partial sum Si = X1 + X2 + + Xi .
We are interested in expansions of the form
k ( )
1
(1) f (M) = Ci (n, f , M) + o
(n)
k+1
,
i=0 n 2
uniformly in M, for k 0. The well-known classical Edgeworth expansions are of the
form in (1) (see Section 2, below, as well as for instance Gnedenko and Kolmogorov
(1954, Chapters 8 and 9) and Petrov (1975, Chapter 7)).
The main idea, here, is to make efficient use of possibly available direct information
on f (i) for small values of i in approximating { f (n) }n1 . In particular, for 0 h n,
(n)
suppose h = h is the probability function for
(2) Sn = Sn,h = X1 + + Xh + Nh+1 + + Nn = Sh + Nh+1 + + Nn ,
where the Ni are i.i.d. (normal) N( , 2 ) random variables, independent of the Xi . Note
that h is the probability function for an independent sum of Sh = X1 + X2 + + Xh
and a normal random variable with mean (n h) and variance (n h) 2 . Assuming
knowledge of the distribution of Sh , h (M), for fixed M, may potentially be computed via
conditioning. In fact, one may make use of existing work regarding sums of particular
random variables with Gaussian random variables (see for instance Nason (2006) for
discussion of the value of results in that direction).
Now, for convenience of notation, set
( )
n k (M) 0 (M)
(3) k (M) = ,
k 0 (M)
( ) ( ) ( )
1 1 (M) 0 (M) 1
(9) fn (M) = 1 (M) + o =n + o 1/2 ,
n1/2 0 (M) n
that is, since 0 (M) = O(1/ n), employing (4),
( )
1
fn (M) = n (1 (M) 0 (M)) + 0 (M) + o
n
( )
1
(10) = n1 (M) (n 1)0 (M) + o .
n
( )
1
fn (M) = 1 (M) + (n 1)(2 (M) 1 (M)) + o
n
( )
1 (M) 0 (M)
= n
0 (M)
( ( ) ( ))
n 2 (M) 0 (M) 1 (M) 0 (M)
+(n 1) n
2 0 (M) 0 (M)
( )
1
+o ,
n
and thus,
n(n 1)
fn (M) = 0 (M) + n(1 (M) 0 (M)) + (2 (M) 0 (M))
(2 )
1
n(n 1)(1 (M) 0 (M)) + o 3/2
n
( )
n(n 1) (n 1)(n 2) 1
= 2 (M) n(n 2)1 (M) + 0 (M) + o 3/2 .
2 2 n
(11)
4 K. S. BERENHAUT, J. W. CHERNESKY, JR., AND R. P. HILTON
Now, combining (7) and (10) gives
( )
( ) 1
fn (M) = n pgn1,p (M 1) + (1 p)gn1,p (M) (n 1)gn,p (M) + o
n
( )2
np 12
M1(n1)p
= e (n1)p(1p)
2 (n 1)p(1 p)
( )2
n(1 p) 12 M(n1)p
+ e (n1)p(1p)
2 (n 1)p(1 p)
( )2
( )
n1 12 Mnp 1
e np(1p)
+o
2 np(1 p) n
( )2
1 M1(n1)p ( )2
n M(n1)p
pe 2 (n1)p(1p) + (1 p)e 2 n1
1
=
2 (n 1)p(1 p)
( )2
( )
n1 12 Mnp 1
(12) e np(1p)
+o .
2 np(1 p) n
The estimate in (12) may be compared with the first-order Edgeworth estimate (see
Proposition 2, below)
( ) ( )3
1 1 2p M np
fn (M) = 0 (M) 1 +
n 6 p(1 p) np(1 p)
( ))) ( )
M np 1
3 +o
np(1 p) n
( )2
1 12 Mnp
= e np(1p)
2 np(1 p)
( ) ( )3 ( )
1 1 2p M np M np
1 + 3
n 6 p(1 p) np(1 p) np(1 p)
( )
1
(13) +o .
n
A simplified version of the estimate in (12), for (n, p) = (50, 3/5) is given by
( )
3 (5M152)2
588
(5M147)2
588
(M30)2
f50 (M) 300e + 200e 343 2e 24 ,
84
(14)
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 5
while the corresponding estimate in (13) is
6 (M30)2
(15) f50 (M) e 24 (M 3 + 90M 2 2664M + 30240).
51840
The differences between the estimates and targets in (14) and (15) are plotted in Figure
1, below. Note the overall marked improvement of the estimate in (14) over that in (15).
F IGURE 1. Line plots of the differences between the estimates and tar-
gets in (14) (dashed) and (15) (solid).
Due to its simplicity, it is worth stating the lower-order expansion in (10) as a formal
result (see Theorem 2, below for the more general expansion in terms of fn ). Note that
beyond the mean and variance of X, for a single value of M, the estimate in (17) requires
only computation of one convolution integral (for further discussion of computational
coinsiderations see Examples 3, 4 and 5, below).
6 K. S. BERENHAUT, J. W. CHERNESKY, JR., AND R. P. HILTON
Proof. This follows from setting k = 1 in Theorem 2, below, and employing the trans-
formations in (4) and (3).
Note that Theorem 4, below, implies that if, in addition, X1 is a symmetric random
variable satisfying E(|X1 |7 ) < , then the o(1/n) error term, in (17) may be replaced
with o(1/n2 ).
For ease of notation, in regards to lattice random variables, we will restrict attention
to those with maximal span h = 1. It is not difficult to modify the given results to other
values of h.
We now turn to a simple continuous random variable example.
Example 2. Consider the case that X1 is a uniform random variable on the interval (0, 1).
In this case (see for instance Mood, Graybill, and Boes (1974, p. 238)),
( )
1 j n
(n 1)! 0
(1) (M j)n1 , If 0 < M < n,
(18) fn (M) = j .
jM
0, otherwise
Theorem 1 gives
( )
1
fn (M) = n1 (M) (n 1)0 (M) + o
n
( )
1
(19) = n(n1 (M) n1 (M 1)) (n 1)n (M) + o ,
n
where i is the cumulative distribution function of a N(i/2, i/12) random variable, and
i is the associated density.
1X is said to be a lattice random variable if there exist constants c and h > 0 such that X takes on
1
values of the form c ih, i Z, with probability one. The constant h is called a span of the distribution.
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 7
Since the distribution of X1 is symmetric, the first-order Edgeworth estimate in this
case is simply n (M) = 0 (M) (see Proposition 2 below). In Section 4, below (see
also the comment following Theorem 1), it is shown that the expansion in (6) can have
distinct advantages in regards to accuracy in the symmetric random variables case.
The following figure contains plots of the differences (n1 (M) (n 1)0 (M))
fn (M) and (the in general much larger in absolute size) 0 (M) fn (M), for n = 100.
F IGURE 2. Plots of the differences 0 (M) fn (M) (left) and (n1 (M)
(n 1)0 (M)) fn (M) (right) for n = 100 and X1 uniform on (0, 1).
We now turn to a comparison of the estimate in (17) with the standard saddlepoint
approximation (see, for instance Goutis and Casella (1999) and the references therein).
i.e. the result of applying Stirlings approximation for the factorials in the binomial
probability function.
8 K. S. BERENHAUT, J. W. CHERNESKY, JR., AND R. P. HILTON
Figure 3 gives the normal, first-order Edgeworth, saddlepoint and standardized sad-
dlepoint approximations (where the standardized saddlepoint is standardized to sum to
one), along with the approximation from Theorem 1, for n = 20 and p = 0.6.
n
0.004
t Theorem 1
s Saddlepoint
d Saddlepoint (st) n
e Edgeworth s
e
n s
0.002
n Normal s s
s e s
Difference from target
e n n
s s
s t n
n s e
0.000
e e e s t t dt d d d st e n
e
st
n
d e
st
n
d e
st
n
d e
st
n
d st
d s
d st
d et
d t et d st
d e
st
n nt d d dt dt d
d
t t
d
n e n
e e
e
n n
e
0.002
n e
n
0.004
0 5 10 15 20
F IGURE 3. Line plots of the differences between the estimates and tar-
gets for n = 20 and p = 0.5 in Example 3.
Table 1 gives the absolute error and rank placement for each of the approximations,
for the 21 values in the support set of Sn ; the respective mean absolute errors (MAE)
are given in the last row of the table. The approximation in (17) fares quite well through
much of the support set. For the corresponding symmetric case (p = 0.5 and n = 20), the
mean absolute errors are given in Table 2. A corresponding plot is given in Figure 4; note
that in this case the Edgeworth and normal approximations coincide (only the Edgeworth
is shown). The estimate in Theorem 1 outperforms the saddlepoint, marginally, with
respect to mean absolute error, in this case.
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 9
TABLE 1. Absolute errors, rank placements and mean absolute errors
for Example 3; as in Figure 3, t, s, d, e and n denote the estimates in
the case of Theorem 1, the saddlepoint, the standardized saddlepoint, the
Edgeworth, and the normal approximation, respectively.
M Thm 1 Saddlept Saddlept(st) Edgew Normal t s d e n
0 0.000000 0.000000 0.000000 0.000000 0.000000 1 2 2 5 4
1 0.000000 0.000000 0.000000 0.000001 0.000000 3 2 1 5 4
2 0.000000 0.000000 0.000000 0.000007 0.000001 3 2 1 5 4
3 0.000000 0.000001 0.000001 0.000031 0.000003 1 3 2 5 4
4 0.000006 0.000006 0.000002 0.000095 0.000038 2 3 1 5 4
5 0.000018 0.000024 0.000005 0.000198 0.000189 2 3 1 5 4
6 0.000012 0.000077 0.000006 0.000231 0.000572 2 3 1 4 5
7 0.000057 0.000207 0.000004 0.000062 0.001095 2 4 1 3 5
8 0.000141 0.000471 0.000044 0.000786 0.001105 2 3 1 4 5
9 0.000030 0.000905 0.000125 0.001358 0.000313 1 4 2 5 3
10 0.000285 0.001473 0.000226 0.000712 0.002901 2 4 1 3 5
11 0.000325 0.002036 0.000282 0.001159 0.004340 2 4 1 3 5
12 0.000158 0.002386 0.000223 0.002386 0.002386 1 3 2 3 5
13 0.000507 0.002362 0.000048 0.001377 0.001804 2 5 1 3 4
14 0.000170 0.001964 0.000153 0.000757 0.004370 2 4 1 3 5
15 0.000308 0.001358 0.000269 0.001668 0.003339 2 3 1 4 5
16 0.000278 0.000772 0.000260 0.000917 0.000598 2 4 1 5 3
17 0.000020 0.000356 0.000174 0.000086 0.001119 1 4 3 2 5
18 0.000128 0.000132 0.000086 0.000392 0.001195 2 3 1 4 5
19 0.000054 0.000041 0.000034 0.000231 0.000618 3 2 1 4 5
20 0.000008 0.000037 0.000037 0.000062 0.000195 1 2 2 4 5
MAE 0.000119 0.000696 0.000094 0.000596 0.001247
e
s
0.002
t Theorem 1 s s
s Saddlepoint
d Saddlepoint (st)
e Edgeworth s s
e e
0.001
s s
Difference from target
s s
e e s s e e
e s t t s e
0.000
e
st
d e
st
d st
d st
d dt d d dt dt d d dt st
d st
d e
st
d e
st
d
t t d t t d t t
e d dt d e
e e
0.001
e e
e e
0 5 10 15 20
F IGURE 4. Line plots of the differences between the estimates and tar-
gets for n = 20 and p = 0.5 in Example 3.
10 K. S. BERENHAUT, J. W. CHERNESKY, JR., AND R. P. HILTON
TABLE 2. Mean absolute errors for n = 20 and p = 0.5 in Example 3.
Note that, formally, the standard saddlepoint approximation requires inverting of the
cumulant generating function, whereas the approximation presented in Theorem 1 is,
in a sense, cumulant-free, but, in place, requires knowledge of the probability function
f . For discussion of saddlepoint approximations in the case of intractable cumulant
generating functions see Kolassa (1991).
In the next and final example of this section, we consider a case where a closed form
for the probability function of Sn is not as readily available. In order to be able to com-
pare estimates, we make use of the valuable distr and distrEx packages in R (see
Ruckdeschel, Kohl, Stabla, and Camphausen (2006)). The packages allow for distribu-
tional computations, for arbitrary probability functions and moderately sized n.
( + ) 1
f (x) = x (1 x) 1
( )( )
(1.2)
(23) = x0.5 (1 x)0.3 , 0 x 1,
(0.5)(0.7)
where is the gamma function given by (z) = 0 et t z1 dt, for z > 0.
Note that here the cumulant generating function is given by log (1 F1 ( ; + ;t)),
where 1 F1 is the confluent hypergeometric function, and the standard saddlepoint ap-
proximation is not as readily available (see Butler (2007, p .16) for discussion). For
some work on densities of sums of Beta random variables, see for instance Pham-Gia
and Turkkan (1998).
Figure 5 provides a comparison of performance of the normal approximation, first-
order Edgeworth approximation and the approximation given in Theorem 1. The approx-
imate mean absolute errors are 0.000934, 0.000420 and 0.000068, respectively. Again
the estimate in (17) performs quite well throughout the support set.
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 11
0.004
n t Theorem 1
e Edgeworth
n
n Normal
ee
0.002
n
n
0.004
n
n
0 5 10 15 20
F IGURE 5. Line plots of the differences between the estimates and tar-
gets for = 0.5 and = 0.7 in Example 4.
As suggested earlier, to obtain the estimate in Theorem 1, for a given X1 with density
f , mean and variance 2 , at a given point, M (and any n), requires numeric evaluation
of only the integral
(24) I(M) = f (x)n1 (M x)dx
where i , here, is the density of a normal random variable with mean i and variance
i 2 . The estimate for fn (M) is then given by
For fixed j, the first four { j,k } are given by j,1 = 0, j,2 = j,2 = 2 , j,3 = j,3
and j,4 = j,4 3 2j,2 (see for instance Gnedenko and Kolmogorov (1954, Chapter 2;
Section 15)).
The Hermite polynomials {Hk } can be defined via
k x2 /2 d k x2 /2
(28) Hk (x) = (1) e e ,
dxk
and satisfy
[ 2k ]
(1) j xk2 j
(29) Hk (x) = k! ,
j=0 j!(k 2 j)!2 j
for k 0. The first six Hermite polynomials are H0 (x) = 1, H1 (x) = x, H2 (x) = x2 1,
H3 (x) = x3 3x, H4 (x) = x4 6x2 + 3, and H5 (x) = x5 10x3 + 15x. Note that these
polynomials satisfy the recurrence relation
(30) Hi (x) = xHi1 (x) (i 1)Hi2 (x),
for i 2. For further information on Hermite polynomials see for instance Szego (1975,
Chapter 5).
Next, define the standardized average cumulants
( j )
1
i=1 i,k
(31) j,k = k
k! j
for j, k 1 and the polynomials
a a
1 ( )k
(32) qa,n (x) = Ha+2v(x) km! n,m+2 m , a, n 1
v=1 k m=1
where the second sum in (32) is over a-tuples k = (k1 , k2 , . . . , ka ) where ki 0 and
ai=1 iki = a, and ai=1 ki = v. For convenience, unless otherwise stated, all sums over k,
throughout, will refer to sums over such a-tuples.
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 13
In addition, set V j (x) = P(Y j < x) and v j (t) = E(eitY j ). Note that if {Y j } are identically
distributed, (32) becomes
a a
1 ( )km
(33) qa,n (x) = qa (x) = Ha+2v(x) km!
m+2 , a, n 1
v=1 k m=1
where for k 1,
k
(34) k = ,
k k!
with k = 1,k .
Now, consider the partial sum Sn = ni=1 Yk for n 1 and note that E[Sn ] = n and
Var(Sk ) = n 2 . Petrov (see Petrov (1975, Theorem 7.14; p. 206)) provides the following
result.
Proposition 1. Suppose {Y j } have zero mean and common finite variance 2 ,
n
1
(35) lim sup
n E(|Y jk+2|) < ,
j=1
and
n
1
(36)
n
|x|k+2 dV j (x) 0,
j=1 |x|>n
for some integer k 1 and some positive < 1/2. In addition, suppose that
n ( )
1
(37) |v j (t)|dt = o n(k+1)/2
|t|> j=1
for every fixed > 0. Then for all sufficiently large n there exists an everywhere contin-
uous density rn (x) of the variable
1
(38) Sn
n
and
( )
1 x2 /2 1 x2 /2 k qa,n (x) 1
(39) rn (x) = e
2
+ e
2
a/2 + o nk/2
a=1 n
uniformly in x ( < x < ).
In the case of identically distributed random variables, Petrov (see Petrov (1975, The-
orems 7.13 and 7.15)) also proves the following result.
Proposition 2. Suppose X1 satisfies E(|X1 |k+2 ) < for some k 1 and X1 satisfies
either
1
Sn has a bounded density rN for some n = N
n
14 K. S. BERENHAUT, J. W. CHERNESKY, JR., AND R. P. HILTON
or
X1 is integer valued with maximal span equal to one,
then, Equation (5) holds with
qi (x)
(40) Ci (n, f , M) =
=: Ei (M), i 1,
ni/2
say, uniformly in < M < , where
M n
(41) x= .
n
Some lower-order Edgeworth approximations implied by Proposition 2, are the fol-
lowing:
( )
1 1
fn (M)
= H3 (x)3 + o ,
n n
( ) ( )
1 1 H6 (x) 2 1
fn (M) = H3 (x)3 + (3 ) + H4 (x)4 + o ,
n n 2! n
( )
1 1 H6 (x) 2
fn (M) = H3 (x)3 + (3 ) + H4 (x)4
n n 2!
( ) ( )
1 H9 (x) 3 1
+ 3/2 (3 ) + H7 (x)3 4 + H5 (x)5 + o 3/2 .
n 3! n
(42)
In the next section, we prove our main result.
3. M AIN RESULT
In this section, we employ Propositions 1 and 2 to prove our main result. While the
results that follow are quite straightforward to state, some care must be taken in tracking
error terms.
Note that throughout, will denote the standard forward difference operator (see, for
instance, Riordan (1979, p. 201)), i.e. for a 0 and given function ,
a ( )
ai a
(43) (x) = (1)
a
(x + i).
i=0 i
The following elementary combinatorial identity will be helpful.
Lemma 1. Suppose k, v 1. Then,
( )
k ( ) 1
1 n1 O vk = o(1), If v > k
(44)
nv1 j=1 j 1
j1 (1)v1 =
n ,
1, otherwise
where j1 (1)v1 is an abbreviation for j1 xv1 evaluated at x = 1.
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 15
Proof. First, note that standard identities for binomial coefficients and the Stirling num-
bers of the second kind {Sn,k } (see for instance Charalambides and Singh (1988), Guo
(2005) and Riordan (1979, p. 202)) give that
j1 (1)v1 = j1 (0)v1 + j (0)v1
= ( j 1)!Sv1, j1 + j!Sv1, j
= ( j 1)!(Sv1, j1 + jSv1, j ) = ( j 1)!Sv, j ,
and
v
(45) (n) j Sv, j = nv,
j=1
Proof. We apply Proposition 1. Since E(|Ni | j ) < for all i h + 1 and j 0 and
E(|X k+2 |) < , it suffices to verify that (37) holds. This follows directly from the fact
that almost all Yi are Gaussian. In particular, |vi (t)| 1 for 1 i h and |vi (t)| = e 2 t
1 2 2
(52) CeKn ,
for constants C, K > 0 (depending on and h).
Since it follows from (27) that all cumulants of order greater than two for a normal
random variable are zero (see for instance Davison (2003, p. 45)), n,m+2 in (32) be-
comes
( n )
1 i=1 i,m+2 h
(53) n,m+2 = m+2 = m+2 ,
(m + 2)! n n
and the theorem follows.
am=1 km
Note that for a-tuples k = (k1 , k2 , . . . , ka ) in (51), = v, and hence Ca,v in (51)
is comprised of terms of order v in the {i }.
We are now in a position to state and prove our main theorem. Note that all results,
from here forward, are stated in terms { fn } and {i } (see Equations (4) and (3) for the
simple transformations to { f (n) } and {i }, respectively).
Theorem 2. Suppose X1 satisfies E(|X1 |3k+2 ) < , for some k 1, and either
1
Sn has a bounded density rN for some n = N
n
or
X1 is integer valued with maximal span equal to one,
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 17
then, Equation (5) holds with
( )
n 1 ( i1 )
(54) Ci (n, f , M) = 1 (M) =: Vi (M), i 1,
i1
( )
k
n 1 ( j1 )
1 (M)
j=1 j 1
( )
3k ( ) ( ) ( k )
1 a a 1 k n1 n
= n Ca,v nv1 j 1 (1) j1 v1
+ o 3k/2
n
a=1 v=1 j=1
( )
k ( ) 3k ( ) ( )
1 a a 1 a k 1
= n Ca,v + n Ca,v + o(1) + o nk/2
a=1 v=1 a=k+1 v=1
k ( )a a ( )
1 1
= n Ca,v + o nk/2
a=1 v=1
( )
1
(56) = fn (M) + o k/2 .
n
The main error term for the kth-order estimate in (54) can actually be computed quite
simply (assuming non-zero third cumulant), as can be seen by the next theorem.
Theorem 3. Suppose the assumptions of Theorem 2 are satisfied, and in addition E(|X1 |3k+3 ) <
. Then,
( )k+1 ( )
k
1 (3 )k+1 1
(57) fn (M) Vi(M) =
n
H3(k+1) (x)
(k + 1)!
+ o (k+1)/2 ,
n
i=1
( )
3k+1 ( )a ( )
k
1 a
1 k
n1
V j (M) = Ca,v (1)
j1 v1
j=1 a=1 n v=1 nv1 j=1 j 1
( )
nk
+o
n(3k+1)/2
( ) 3k+1 ( ) ( )
k
1 a a 1 a k
= Ca,v + n Ca,v + o(1)
a=1 n v=1 a=k+2 v=1
( )k+1 ( k ) ( )
1 1
+
n Ck+1,v + o(1) + o n(k+1)/2
v=1
( ) ( )
1 k+1 1
(58) = fn (M) Ck+1,k+1 + o (k+1)/2 .
n n
In the next section we will discuss some implications of Theorem 2.
4. S OME I MPLICATIONS
In this section we will consider some implications of Theorem 2, including in partic-
ular, improved errors in the case of symmetric distributions as well as facility in deter-
mining asymptotic crossing points of estimates with the target.
k ( )
1
(59) fn (M) = Vi(M) + o n(2k+1)/2
,
i=1
Theorem 5. Suppose that X1 is a symmetric integer valued random variable with maxi-
mal span equal to one and satisfying E(|X1 |2k/2+3 ) < for some k 1. Then,
k ( )
1
(60) fn (M) = Ei(M) + o n(2k/2+1)/2
,
i=1
k 1 2 3 4 5
Edgeworth Expansion 1/n 1/2 1/n 3/2 1/n 3/2 1/n5/2 1/n5/2
Expansion given in (54) 1/n3/2 1/n5/2 1/n7/2 1/n9/2 1/n11/2
Arguing similar to in the proof of Theorem 3, we have the following regarding the
main error term in the case of symmetric distributions.
Theorem 6. Suppose the assumptions of Theorem 4 are satisfied, and in addition E(|X1 |4k+5 ) <
, for some integer k 1. Then,
( )
k
1 (4 )k+1 1
(61) fn (M) Vi (M) = k+1 H4(k+1) (x)
+ o (2k+3)/2
i=1 n (k + 1)! n
uniformly in < M < , where x is as in (41).
4.2. Asymptotic crossing points. Asymptotic crossing points between the kth-order
estimate in (54) and the target probability function can be computed quite simply as can
be seen by the next theorem. Note that the result, and discussion that follows, is in terms
of unstandardized random variables; the convergent nature of the crossing points is, of
course, in standardized terms.
Theorem 7. Suppose the assumptions of Theorem 3 are satisfied. Then, if 3 = 0, the
(unstandardized) asymptotic crossing points of the estimate with the target are
(62) i n + n , 1 i 3(k + 1),
where 1 < 2 < < 3(k+1) are the 3(k + 1) real and simple zeroes of the Hermite
polynomial H3(k+1) .
Proof. This follows directly from Theorem 3 and the fact that all zeroes of Hermite
polynomials are real and simple (see for instance Szego (1975, Section 6.2)).
Note that without additional assumptions the main error term in (40) is simply
( )
1 k+1
(63) qk+1 (x),
n
ASYMPTOTIC EXPANSIONS VIA LOWER-ORDER CONVOLUTIONS 21
i.e. a potentially several-term linear combination of Hermite polynomials in x. In this
case, the many results on the locations of zeroes of Hermite polynomials (see for instance
Szego (1975, Chapter 6)) cannot be directly applied. In fact the zeroes of qk+1 (x) need
not be real as is seen by the following example.
to above, the asymptotic crossing points of the estimate with the target are M =
Similar
i n + n , where i are the distinct real zeros of the polynomial Z defined via
(3 ) 3
(64) Z(x) = H9 (x) + H7 (x)4 3 + H5 (x)5 .
3!
Note that the zeroes of Z, are dependent on the distribution of X1 through the cumulants
3 , 4 and 5 as well as the variance 2 . In the case where X is supported on {0, 1, 2}
with probabilities P(X = 0) = 7/22, P(X = 1) = 14/22 and P(X = 2) = 1/22, we have
( 2 , 3 , 4 , 5 ) = (35/121, 21/1331, 574/14641, 21525/161051)
and
35
(65) Z(x) = y(417255 + 263190y2 21762y4 446y6 + y8 ).
7350000
There are seven real zeros of Z, namely
(66) x 0, 1.375929290, 2.897073323, 22.12179176,
and two complex zeros
(67) x 7.325291i.
Note that the analogous result to Theorem 7 for symmetric distributions follows from
Theorem 6.
22 K. S. BERENHAUT, J. W. CHERNESKY, JR., AND R. P. HILTON
For further work on asymptotic crossing points see for instance Finner, Roters, and
Dickhaus (2007) and Berenhaut, Chen, and Tran (2008) and the references therein.
Acknowledgements
We are very thankful to a referee for comments and insights that improved this man-
uscript.
R EFERENCES
Berenhaut, K. S., Chen, D., & Tran, V. (2008). On the compatibility of Dysons condi-
tions. Statist. Probab. Lett., 78(17), 31103113.
Butler, R. W. (2007). Saddlepoint approximations with applications. Cambridge: Cam-
bridge University Press.
Charalambides, C. A., & Singh, J. (1988). A review of the Stirling numbers, their
generalizations and statistical applications. Comm. Statist. Theory Methods, 17(8),
25332595.
Churchill, E. (1946). Information given by odd moments. Ann. Math. Statistics, 17,
244246.
Davison, A. C. (2003). Statistical models (Vol. 11). Cambridge: Cambridge University
Press.
Finner, H., Roters, M., & Dickhaus, T. (2007). Characterizing density crossing points.
Amer. Statist., 61(1), 2833.
Gnedenko, B. V., & Kolmogorov, A. N. (1954). Limit distributions for sums of inde-
pendent random variables. Inc., Cambridge, Mass.: Addison-Wesley Publishing
Company. (Translated and annotated by K. L. Chung. With an Appendix by J. L.
Doob)
Goutis, C., & Casella, G. (1999). Explaining the saddlepoint approximation. Amer.
Statist., 53(3), 216224.
Guo, L. (2005). Baxter algebras, Stirling numbers and partitions. J. Algebra Appl., 4(2),
153164.
Kolassa, J. E. (1991). Saddlepoint approximations in the case of intractable cumulant
generating functions. In Selected Proceedings of the Sheffield Symposium on Ap-
plied Probability (Sheffield, 1989) (Vol. 18, pp. 236255). Hayward, CA: Inst.
Math. Statist.
Mood, A., Graybill, F., & Boes, D. (1974). Introduction to the theory of statistics (3. ed.
ed.). Auckland [u.a.]: McGraw-Hill.
Nason, G. P. (2006). On the sum of t and Gaussian random variables. Statist. Probab.
Lett., 76(12), 12801286.
Petrov, V. V. (1975). Sums of independent random variables. New York: Springer-
Verlag. (Translated from the Russian by A. A. Brown, Ergebnisse der Mathematik
und ihrer Grenzgebiete, Band 82)
Pham-Gia, T., & Turkkan, N. (1998). Distribution of the linear combination of two
general beta variables and applications. Comm. Statist. Theory Methods, 27(7),
18511869.
References 23
Riordan, J. (1979). Combinatorial identities. Robert E. Krieger Publishing Co., Hunt-
ington, N.Y. (Reprint of the 1968 original)
Ruckdeschel, P., Kohl, M., Stabla, T., & Camphausen, F. (2006, May). S4 classes for
distributions. R News, 6(2), 26.
Szego, G. (1975). Orthogonal polynomials (Fourth ed.). Providence, R.I.: American
Mathematical Society. (American Mathematical Society, Colloquium Publica-
tions, Vol. XXIII)
D EPARTMENT OF M ATHEMATICS , WAKE F OREST U NIVERSITY, W INSTON -S ALEM , NC 27109
E-mail address: berenhks@wfu.edu