1 Introduction

The classical Stirling numbers play an important role in many branches of mathematics and physics as ingredients in the computation of diverse quantities. In particular, the Stirling numbers of the second kind S(jm), counting the number of partitions of \(\{1,\ldots ,j\}\) into m nonempty, pairwise disjoint subsets, are a fundamental tool in many combinatorial problems. Such numbers can be defined in various equivalent ways (cf. Abramowitz and Stegun [1, p. 824] and Comtet [7, Chap. 5]). Two of the most useful are the following. Let \(j=0,1,\ldots \) and \(m=0,1,\ldots ,j\). Then S(jm) can be explicitly defined as

$$\begin{aligned} S(j,m)=\dfrac{1}{m!}\sum _{k=0}^m \left( {\begin{array}{c}m\\ k\end{array}}\right) (-1)^{m-k}k^j, \end{aligned}$$
(1)

or via their generating function as

$$\begin{aligned} \dfrac{(e^z-1)^m}{m!}=\sum _{j=m}^\infty S(j,m)\dfrac{z^j}{j!},\quad z\in \mathbb {C}. \end{aligned}$$
(2)

Motivated by various specific problems, different generalizations of the Stirling numbers S(jm) have been considered in the literature (see, for instance, Hsu and Shiue [11], Luo and Srivastava [13], Cakić et al. [6] and El-Desouky et al. [9]). In [3], we considered the following probabilistic generalization. Let \((Y_k)_{k\ge 1}\) be a sequence of independent copies of a real-valued random variable Y having a finite moment generating function and denote by \(S_k=Y_1+\cdots +Y_k, k=1,2,\ldots \) \((S_0=0)\). Then, the Stirling numbers of the second kind associated with Y are defined by

$$\begin{aligned} S_Y(j,m)=\dfrac{1}{m!}\sum _{k=0}^m \left( {\begin{array}{c}m\\ k\end{array}}\right) (-1)^{m-k}\mathbb {E}S_k^j,\quad j=0,1,\ldots ,\quad m=0,1,\ldots ,j. \end{aligned}$$
(3)

Observe that formula (3) recovers (1) when \(Y=1\). The motivations behind definition (3) have to do with certain problems coming from analytic number theory, such as extensions in various ways of the classical formula for sums of powers on arithmetic progressions (cf. [3]) and explicit expressions for higher-order convolutions of Appell polynomials (see [4]).

In this paper, we extend definition (3) to complex-valued random variables Y and show its usefulness in various classical topics of probability theory. In this regard, we show in Sect. 3 that the moments \(\mathbb {E}S_n^j\) can be written in closed form in terms of the Stirling numbers \(S_Y(j,m)\). When Y is real-valued and centered, two remarkable consequences deserve to be mentioned. First, we can directly obtain the precise asymptotic behavior of \(\mathbb {E}S_n^j\) as far as rates of convergence and leading coefficients are concerned, without appealing to the central limit theorem. Monotonicity properties of the sequence \(\mathbb {E}S_n^{2j}/(n)_j, n\ge j\), where \((n)_j=n(n-1)\cdots (n-j+1)\), are also derived in a simple way. We point out that monotonicity results in the central limit theorem seem to be rather scarce. (In this respect, Teicher [19] and Kane [12] showed that \(P(S_n\ge 0)\) converges monotonically for various choices of the law of \(S_n\).) Second, from a computational point of view, we can evaluate \(\mathbb {E}S_n^j\) for \(n\ge j\) in terms of \(\mathbb {E}S_n^j\) for \(n<j\) and \(S_Y(j,\lfloor j/2\rfloor )\). In Sect. 4, we deal with analogous properties referring to Lévy processes and centered subordinators.

Concerning rates of convergence in the central limit theorem, Edgeworth expansions provide a great accuracy in the approximation at the price of using rather involved technicalities (cf. Petrov [15], Hall [10], Barbour [5], and Rinot and Rotar [16]). In Sect. 5, we give explicit and relatively simple full Edgeworth expansions whose coefficients depend on the Stirling numbers \(S_{Y+iZ}(j,m)\), where the real-valued random variables Y and Z are independent and Z has the standard normal distribution. The order of magnitude of such expansions is that of \(n^{-(r-1)/2}\), whenever \(\mathbb {E}Y^k=\mathbb {E}Z^k, k=1,2,\ldots ,r\), for some \(r\ge 2\). In Sect. 6, we show that the cumulants of a random variable Y can also be described by means of \(S_Y(j,m)\). Finally, in Sect. 2 we gather some equivalent definitions of \(S_Y(j,m)\) when Y is complex-valued without proofs, since they are similar to those previously given in [3] for real-valued random variables Y.

2 Probabilistic Stirling Numbers

The following notations will be used throughout the paper. Let \(\mathbb {N}\) be the set of positive integers and \(\mathbb {N}_0=\mathbb {N}\cup \{0\}\). Unless otherwise specified, we assume that \(j,m\in \mathbb {N}_0\), \(x\in \mathbb {C}\), and \(z\in \mathbb {C}\) satisfies \(|z|<R\), where \(R>0\) may change from line to line. We always consider measurable exponentially bounded functions \(f:\mathbb {C}\rightarrow \mathbb {C}\), i.e., \(|f(x)|\le e^{R|x|}\). We denote by \(I_j(x)=x^j\) the jth monomial function and by \((x)_j\) the descending factorial, that is, \((x)_j=x(x-1)\cdot \cdot \cdot (x-j+1), j\in \mathbb {N}\), \((x)_0=1\). Finally, we set \(j\wedge m= \min (j,m)\) and denote by \(\lfloor y\rfloor \) the integer part of \(y\in \mathbb {R}\).

Let \({\mathcal {G}}_0\) be the set of complex-valued random variables Y having a finite moment generating function in a neighborhood of the origin, i.e.,

$$\begin{aligned} \mathbb {E}e^{|zY|}<\infty ,\quad |z|<R, \end{aligned}$$

for some \(R>0\).

For any \(r\in \mathbb {N}\), we consider a random variable \(\beta (r)\) having the beta density

$$\begin{aligned} \rho _r(\theta )=r(1-\theta )^{r-1},\quad 0\le \theta \le 1, \end{aligned}$$
(4)

whereas we set \(\beta (0)=1\). Note that \(\beta (1)\) is uniformly distributed on [0, 1]. For any \(r\in \mathbb {N}_0\), let \((Y_k)_{k\ge 1}\) and \((\beta _k(r))_{k\ge 1}\) be two sequences of independent copies of \(Y\in {\mathcal {G}}_0\) and \(\beta (r)\), respectively, and assume that both sequences are mutually independent. We denote

$$\begin{aligned} W_m(r,Y)= \beta _1(r)Y_1+\cdots +\beta _m(r)Y_m,\quad m\in \mathbb {N}\quad (W_0(r,Y)=0). \end{aligned}$$
(5)

The following two important special cases will also be denoted

$$\begin{aligned} W_m(0,Y)=Y_1+\cdots +Y_m=S_m,\qquad W_m(2,Y)=W_m(Y),\quad m\in \mathbb {N}_0. \end{aligned}$$
(6)

On the other hand, consider the difference operator

$$\begin{aligned} \triangle _y^1f(x)=f(x+y)-f(x),\quad y\in \mathbb {C}, \end{aligned}$$

together with the iterates

$$\begin{aligned} \triangle _{y_1,\ldots ,y_m}^m f(x)=(\triangle _{y_1}^1\circ \cdot \cdot \cdot \circ \triangle _{y_m}^1)f(x),\quad (y_1,\ldots ,y_m)\in \mathbb {C}^m,\quad m\in \mathbb {N}. \end{aligned}$$
(7)

Such generalized difference operators were used by Mrowiec et al. [14] and Dilcher and Vignat [8] in different analytic contexts. Observe that

$$\begin{aligned} \triangle ^m f(x):=\triangle _{1,\ldots ,1}^m f(x)=\sum _{k=0}^m \left( {\begin{array}{c}m\\ k\end{array}}\right) (-1)^{m-k}f(x+k),\quad m\in \mathbb {N}_0, \end{aligned}$$

is the usual mth forward difference of f. In general, the iterates in (7) have a cumbersome expression. However, we have the following formulas stated in [3], where it is understood that

$$\begin{aligned} \triangle _\emptyset ^m f(x)=f(x),\qquad \prod _{k=1}^0=1. \end{aligned}$$

Lemma 1

Let \(Y\in {\mathcal {G}}_0\). For any \(m\in \mathbb {N}_0\), we have

$$\begin{aligned} \mathbb {E}\triangle _{Y_1,\ldots ,Y_m}^m f(x)=\sum _{k=0}^m \left( {\begin{array}{c}m\\ k\end{array}}\right) (-1)^{m-k}\mathbb {E}f(x+S_k). \end{aligned}$$

If, in addition, f is m times differentiable, then

$$\begin{aligned} \mathbb {E}\triangle _{Y_1,\ldots ,Y_m}^m f(x)=\mathbb {E}Y_1\cdot \cdot \cdot Y_m f^{(m)}(x+W_m(1,Y)). \end{aligned}$$

The Stirling numbers of the second kind \(S_Y(j,m), m\le j\), associated with the random variable \(Y\in {\mathcal {G}}_0\) are defined as in (3). Observe that this definition is justified in the sense that

$$\begin{aligned} S_Y(j,m)=0,\quad m>j, \end{aligned}$$
(8)

as follows by choosing \(f=I_j\) and \(x=0\) in Lemma 1. Such numbers are characterized in the following result (cf. [3]).

Theorem 1

Let \(Y\in {\mathcal {G}}_0\). For any \(m\le j\), we have

$$\begin{aligned} \begin{aligned} S_Y(j,m)&=\frac{1}{m!}\mathbb {E}\triangle _{Y_1,\ldots ,Y_m}^m I_j(0)=\left( {\begin{array}{c}j\\ m\end{array}}\right) \mathbb {E}Y_1\cdot \cdot \cdot Y_m W_m(1,Y)^{j-m}\\&= \frac{1}{m!}\sum _{l=0}^j S(j,l)\sum _{k=0}^m \left( {\begin{array}{c}m\\ k\end{array}}\right) (-1)^{m-k}\mathbb {E}(S_k)_l. \end{aligned} \end{aligned}$$
(9)

Equivalently, the numbers \(S_Y(j,m)\) are defined via their generating function as

$$\begin{aligned} \frac{1}{m!}\left( \mathbb {E}e^{zY}-1\right) ^m=\sum _{j=m}^\infty S_Y(j,m)\frac{z^j}{j!}. \end{aligned}$$
(10)

For the classical Stirling numbers S(jm), expression (9) gives us

$$\begin{aligned} S(j,m)=\left( {\begin{array}{c}j\\ m\end{array}}\right) \mathbb {E}(\beta _1(1)+\cdots +\beta _m(1))^{j-m},\quad m\le j. \end{aligned}$$

This formula was already obtained by Sun [18].

Theorem 1 allows us to obtain explicit expressions of \(S_Y(j,m)\) for different choices of the random variable Y (see [3]). In many cases, such numbers are actually real numbers. For instance, if \(Y=U+iV\), where U and V are independent real-valued random variables and V has a real characteristic function (in particular, if \(V=0\) or if V has the standard normal distribution). In fact, let \(t\in \mathbb {R}\) with \(|t|<R\). Since \(\mathbb {E}e^{tY}=\mathbb {E}e^{tU}\mathbb {E}e^{itV}\) is real, we see that both sides in (10) are real when choosing \(z=t\). This shows the claim. Finally, if Y is nonnegative, then \(S_Y(j,m)\) is nonnegative as well, as follows from (5) and (9).

3 Moments

In this section, we give closed-form expressions for the moments of \(S_n\), as defined in (6) in terms of the probabilistic Stirling numbers \(S_Y(j,m)\) and discuss some of their consequences. In this respect, for any \(r\in \mathbb {N}\), denote by \({\mathcal {G}}_r\) the subset of \({\mathcal {G}}_0\) consisting of those random variables \(Y=U+iV\) such that

$$\begin{aligned} \mathbb {E}Y^k=0,\quad k=1,\ldots ,r. \end{aligned}$$
(11)

In the case in which U and V are independent, observe that \(Y\in {\mathcal {G}}_1\), if \(\mathbb {E}U=\mathbb {E}V=0\); \(Y\in {\mathcal {G}}_2\) if, in addition, \(\mathbb {E}U^2=\mathbb {E}V^2\); \(Y\in {\mathcal {G}}_3\) if, moreover, \(\mathbb {E}U^3=\mathbb {E}V^3=0\); \(Y\in {\mathcal {G}}_4\) if, additionally,

$$\begin{aligned} \mathbb {E}U^4-\left( {\begin{array}{c}4\\ 2\end{array}}\right) (\mathbb {E}U^2)^2+\mathbb {E}V^4=0, \end{aligned}$$
(12)

and so on. Also observe that if U and V are independent copies of a random variable having the standard normal distribution, then \(Y\in {\mathcal {G}}_r\), for any \(r\in \mathbb {N}\), since

$$\begin{aligned} \mathbb {E}e^{itY}=1,\quad t\in \mathbb {R}. \end{aligned}$$
(13)

The following auxiliary result will be used in Sect. 5.

Lemma 2

Let \(r\in \mathbb {N}\) and let U and V be two independent real-valued random variables such that \(U\in {\mathcal {G}}_0\) and V has the standard normal distribution. Then, \(Y=U+iV\in {\mathcal {G}}_r\) if and only if \(\mathbb {E}U^k=\mathbb {E}V^k, k=1,\ldots ,r\).

Proof

Let Z be an independent copy of V. By (13), we see that

$$\begin{aligned} \mathbb {E}(Z+iV)^l=0,\quad l\in \mathbb {N}. \end{aligned}$$
(14)

Assume that \(\mathbb {E}U^k=\mathbb {E}V^k, k=1,\ldots ,r\), and let \(s=1,\ldots ,r\). By (14), we have

$$\begin{aligned} \begin{aligned} \mathbb {E}(U+iV)^s&=\sum _{k=0}^s \left( {\begin{array}{c}s\\ k\end{array}}\right) \mathbb {E}U^k\mathbb {E}(iV)^{s-k}\\&=\sum _{k=0}^s \left( {\begin{array}{c}s\\ k\end{array}}\right) \mathbb {E}Z^k\mathbb {E}(iV)^{s-k}=\mathbb {E}(Z+iV)^s=0. \end{aligned} \end{aligned}$$

To show the reverse implication, we use induction on r. For \(r=1\), the result is obviously true. Assume that the result is true for some \(r\in \mathbb {N}\). Let \(Y\in {\mathcal {G}}_{r+1}\subseteq {\mathcal {G}}_r\). By the induction assumption, \(\mathbb {E}U^k=\mathbb {E}V^k=\mathbb {E}Z^k\), \(k=1,\ldots ,r\). We thus have from (14)

$$\begin{aligned} \begin{aligned} 0&=\mathbb {E}(U+iV)^{r+1}=\mathbb {E}U^{r+1}+\sum _{k=0}^r \left( {\begin{array}{c}r+1\\ k\end{array}}\right) \mathbb {E}Z^k\mathbb {E}(iV)^{r+1-k}\\&=\mathbb {E}U^{r+1}+\mathbb {E}(Z+iV)^{r+1}-\mathbb {E}Z^{r+1}=\mathbb {E}U^{r+1}-\mathbb {E}Z^{r+1}=\mathbb {E}U^{r+1}-\mathbb {E}V^{r+1}. \end{aligned} \end{aligned}$$

This shows the reverse implication and completes the proof. \(\square \)

The interesting feature of the random variables Y in the subset \({\mathcal {G}}_r, r\in \mathbb {N}\), is that its corresponding Stirling numbers satisfy \(S_Y(j,m)=0\), for \(j<m(r+1)\), as shown in the following result. This property has remarkable consequences to evaluate the moments \(\mathbb {E}S_n^j\), as seen in the remaining results of this section, as well as to obtain the Edgeworth expansions considered in Sect. 5.

Theorem 2

Let \(Y\in {\mathcal {G}}_r\), for some \(r\in \mathbb {N}_0\). Then,

$$\begin{aligned} S_Y(j,m)=\frac{(m(r+1))!}{m!((r+1)!)^m}\left( {\begin{array}{c}j\\ m(r+1)\end{array}}\right) \mathbb {E}(Y_1\cdot \cdot \cdot Y_m)^{r+1} W_m(r+1,Y)^{j-m(r+1)}, \end{aligned}$$

whenever \(j\ge m(r+1)\), whereas \(S_Y(j,m)=0\), if \(j< m(r+1)\).

Proof

We start with the following identity, which follows from the formula for the remainder term in Taylor’s theorem:

$$\begin{aligned} e^z=\sum _{k=0}^r \frac{z^k}{k!}+\frac{z^{r+1}}{(r+1)!}\mathbb {E}e^{z\beta (r+1)}, \end{aligned}$$

where \(\beta (r+1)\) is the random variable defined in (4). Replacing z by zY in this formula and then taking expectations, we have from (5) and (11)

$$\begin{aligned} \begin{aligned} \frac{1}{m!}\left( \mathbb {E}e^{zY}-1\right) ^m&=\frac{z^{m(r+1)}}{m!((r+1)!)^m}\left( \mathbb {E}Y^{r+1}e^{z\beta (r+1)Y}\right) ^m\\&=\frac{z^{m(r+1)}}{m!((r+1)!)^m}\mathbb {E}(Y_1\cdot \cdot \cdot Y_m)^{r+1}e^{zW_m(r+1,Y)}\\&=\frac{z^{m(r+1)}}{m!((r+1)!)^m}\sum _{k=0}^\infty \mathbb {E}(Y_1\cdot \cdot \cdot Y_m)^{r+1}W_m(r+1,Y)^k\frac{z^k}{k!}. \end{aligned} \end{aligned}$$

Thus, the result follows from (10) with the change \(j=m(r+1)+k\). \(\square \)

Theorem 3

Let \(Y\in {\mathcal {G}}_r\), for some \(r\in \mathbb {N}_0\). Denote by \(\tau _r(j)=\lfloor j/(r+1)\rfloor \). For any \(n\in \mathbb {N}_0\), we have

$$\begin{aligned} \mathbb {E}S_n^j=\sum _{m=0}^{n\wedge \tau _r(j)}S_Y(j,m)(n)_m. \end{aligned}$$
(15)

Moreover, for any \(n\ge \tau _r(j)\), we have

$$\begin{aligned} \begin{aligned} \frac{\mathbb {E}S_n^j}{(n)_{\tau _r(j)}} =&S_Y(j,\tau _r(j))\\&+\frac{1}{(\tau _r(j)-1)!}\sum _{k=0}^{\tau _r(j)-1}\left( {\begin{array}{c}\tau _r(j)-1\\ k\end{array}}\right) \frac{(-1)^{\tau _r(j)-k-1}}{n-k}\mathbb {E}S_k^j. \end{aligned} \end{aligned}$$
(16)

Proof

Note that

$$\begin{aligned} \left( \mathbb {E}e^{zY}\right) ^n=\mathbb {E}e^{zS_n}=\sum _{j=0}^\infty \mathbb {E}S_n^j\frac{z^j}{j!}. \end{aligned}$$
(17)

By (10) and Theorem 2, we see that

$$\begin{aligned} \begin{aligned} \left( \mathbb {E}e^{zY}\right) ^n&=\sum _{m=0}^n (n)_m \frac{\left( \mathbb {E}e^{zY}-1\right) ^m}{m!}=\sum _{m=0}^n (n)_m \sum _{j=m(r+1)}^\infty S_Y(j,m)\frac{z^j}{j!}\\&=\sum _{j=0}^\infty \frac{z^j}{j!}\sum _{m=0}^{n\wedge \tau _r(j) } S_Y(j,m)(n)_m. \end{aligned} \end{aligned}$$

This, together with (17), shows (15). On the other hand, if \(n=\tau _r(j)\), formula (16) directly follows from definition (3). Assume that \(n>\tau _r(j)\). The following combinatorial identity

$$\begin{aligned} \sum _{l=0}^p \left( {\begin{array}{c}s\\ l\end{array}}\right) (-1)^l=\left( {\begin{array}{c}s-1\\ p\end{array}}\right) (-1)^p,\qquad p,s\in \mathbb {N}_0,\quad p\le s-1, \end{aligned}$$

can be easily shown by induction on p. Using (3), (15) and the preceding identity, we have

$$\begin{aligned} \begin{aligned} \mathbb {E}S_n^j&=\sum _{m=0}^{\tau _r(j)}\left( {\begin{array}{c}n\\ m\end{array}}\right) \sum _{k=0}^m \left( {\begin{array}{c}m\\ k\end{array}}\right) (-1)^{m-k}\mathbb {E}S_k^j\\&=\sum _{k=0}^{\tau _r(j)}\left( {\begin{array}{c}n\\ k\end{array}}\right) \mathbb {E}S_k^j\sum _{m=k}^{\tau _r(j)}\left( {\begin{array}{c}n-k\\ m-k\end{array}}\right) (-1)^{m-k}\\&=\sum _{k=0}^{\tau _r(j)}\left( {\begin{array}{c}n\\ k\end{array}}\right) \mathbb {E}S_k^j\left( {\begin{array}{c}n-k-1\\ \tau _r(j)-k\end{array}}\right) (-1)^{\tau _r(j)-k}\\&=\left( {\begin{array}{c}n\\ \tau _r(j)\end{array}}\right) \sum _{k=0}^{\tau _r(j)}\left( {\begin{array}{c}\tau _r(j)\\ k\end{array}}\right) \frac{n-\tau _r(j)}{n-k}(-1)^{\tau _r(j)-k}\mathbb {E}S_k^j. \end{aligned} \end{aligned}$$

This, together with definition (3), shows (16). The proof is complete. \(\square \)

The classical Stirling numbers of the second kind S(jm) can also be defined by means of the equations

$$\begin{aligned} x^j=\sum _{m=0}^j S(j,m)(x)_m,\quad j\in \mathbb {N}_0. \end{aligned}$$
(18)

In this sense, formula (15) may be thought as the probabilistic counterpart of (18).

Corollary 1

Let \(Y\in {\mathcal {G}}_0\). Then,

$$\begin{aligned} |S_Y(j,m)|\le \frac{\mathbb {E}(|Y_1|+\cdots +|Y_m|)^j}{m!},\quad 0\le m \le j. \end{aligned}$$

Proof

Let \(0\le m \le j\). By (5) and the second equality in (9), we see that \(|S_Y(j,m)|\le S_{|Y|}(j,m)\). Applying (15) with \(r=0\), we get

$$\begin{aligned} \mathbb {E}(|Y_1|+\cdots +|Y_m|)^j=\sum _{k=0}^m S_{|Y|}(j,k)(m)_k \ge S_{|Y|}(j,m)m!. \end{aligned}$$

This completes the proof. \(\square \)

This result extends the well-known upper bound for the classical Stirling numbers of the second kind, namely

$$\begin{aligned} S(j,m)\le \frac{m^j}{m!},\quad 0\le m \le j. \end{aligned}$$

The case in which \(Y\in {\mathcal {G}}_1\) is real-valued deserves special attention. First, denote by \(\sigma ^2=\mathbb {E}Y^2\) its variance and define the real-valued random variable \({\widetilde{Y}}\) whose distribution function is given by

$$\begin{aligned} F_{{\widetilde{Y}}}(y)=\dfrac{1}{\sigma ^2}\int _{-\infty }^y x^2 F_Y(dx),\quad y\in \mathbb {R}, \end{aligned}$$

where \(F_Y\) is the distribution function of Y and it is assumed that \(\sigma ^2>0\). Note that for any function f we have

$$\begin{aligned} \sigma ^2\mathbb {E}f({\widetilde{Y}})=\mathbb {E}Y^2f(Y). \end{aligned}$$
(19)

In the trivial case in which \(Y=0\), a.s., we define \({\widetilde{Y}}=0\), a.s., so that formula (19) still holds. Second, consider the random variable \(W_m({\widetilde{Y}})\) as defined in (6). Finally, recall that if Z is a random variable having the standard normal distribution, then

$$\begin{aligned} \mathbb {E}Z^{2m}=\frac{(2m)!}{m!2^m},\quad m\in \mathbb {N}_0. \end{aligned}$$
(20)

With these ingredients, we give the following.

Corollary 2

Let \(Y\in {\mathcal {G}}_1\) be real-valued. Then,

$$\begin{aligned} S_Y(j,m)=\left( {\begin{array}{c}j\\ 2m\end{array}}\right) \mathbb {E}(\sigma Z)^{2m}\mathbb {E}W_m({\widetilde{Y}})^{j-2m}, \end{aligned}$$
(21)

whenever \(j\ge 2m\), whereas \(S_Y(j,m)=0\), for \(j<2m\). In particular,

$$\begin{aligned} S_Y(2j,j)=\mathbb {E}(\sigma Z)^{2j},\qquad S_Y(2j+1,j)=j(2j+1)\mathbb {E}(\sigma Z)^{2j}\frac{\mathbb {E}Y^3}{3\sigma ^2}. \end{aligned}$$
(22)

Proof

The proof of (21) follows along the lines of that of Theorem 2, by taking into account (20) and the fact that

$$\begin{aligned} \mathbb {E}e^{zY}-1=\frac{z^2}{2}\mathbb {E}Y^2e^{z\beta (2)Y}=\frac{\sigma ^2 z^2}{2}\mathbb {E}e^{z\beta (2){\widetilde{Y}}}, \end{aligned}$$

as follows from (19). The identities in (22) are a consequence of (21) and the equalities

$$\begin{aligned} \mathbb {E}\beta (2)=\frac{1}{3}, \qquad \mathbb {E}{\widetilde{Y}}=\frac{\mathbb {E}Y^3}{\sigma ^2}. \end{aligned}$$

The proof is complete. \(\square \)

Corollary 3

Let \(Y\in {\mathcal {G}}_1\) be real-valued. Then,

$$\begin{aligned} \mathbb {E}S_n^j=\sum _{m=0}^{n\wedge \lfloor j/2\rfloor }S_Y(j,m)(n)_m. \end{aligned}$$
(23)

As a consequence, the sequence \(\mathbb {E}S_n^{2j}/(n)_j, n\ge j\), decreases to \(\mathbb {E}(\sigma Z)^{2j}\).

Moreover, for any \(n\ge j\), we have

$$\begin{aligned} \begin{aligned} \frac{\mathbb {E}S_n^j}{(n)_{\lfloor j/2\rfloor }} =&S_Y(j,\lfloor j/2\rfloor )\\&+\frac{1}{(\lfloor j/2\rfloor -1)!}\sum _{k=0}^{\lfloor j/2\rfloor -1}\left( {\begin{array}{c}j-1\\ k\end{array}}\right) \frac{(-1)^{\lfloor j/2\rfloor -k-1}}{n-k}\mathbb {E}S_k^j. \end{aligned} \end{aligned}$$
(24)

Proof

As follows from (21), the Stirling numbers \(S_Y(2j,m)\) are positive. This, together with (22) and (23), implies that the sequence \(\mathbb {E}S_n^{2j}/(n)_j, n\ge j\), decreases to \(S_Y(2j,j)=\mathbb {E}(\sigma Z)^{2j}\). The remaining assertions readily follow from Theorem 3 by choosing \(r=1\). The proof is complete. \(\square \)

Let \(Y\in {\mathcal {G}}_1\) be real-valued. Traditionally, the problem of convergence concerning the moments \(\mathbb {E}S_n^j\), as \(n\rightarrow \infty \), is carried out by establishing first the central limit theorem

$$\begin{aligned} \frac{S_n}{\sigma \sqrt{n}}\rightarrow Z,\quad n\rightarrow \infty , \end{aligned}$$

and afterward showing (see, for instance, von Bahr [20]) that

$$\begin{aligned} \mathbb {E}\left( \frac{S_n}{\sigma \sqrt{n}}\right) ^j\rightarrow \mathbb {E}Z^j,\quad n\rightarrow \infty . \end{aligned}$$

The explicit expressions in Corollaries 2 and 3 directly give us the precise asymptotic behaviour of the moments \(\mathbb {E}S_n^j\) as far as rates of convergence and leading coefficients are concerned. Note, in particular, that the odd moments \(\mathbb {E}S_n^{2j+1}\) have the order of magnitude of \((n)_j\) (resp. \((n)_{j-1}\)) with leading coefficients \(S_Y(2j+1,j)\) (resp. \(S_Y(2j+1,j-1)\)) in the case that \(\mathbb {E}Y^3\ne 0\) (resp. \(\mathbb {E}Y^3=0\)), as follows from (22).

On the other hand, \(\mathbb {E}S_n^{2j}/(n)_j\) decreasingly converges to \(\mathbb {E}(\sigma Z)^{2j}\). This monotonicity property is no longer true, in general, for the odd moments, since the leading coefficient of \(\mathbb {E}S_n^{2j+1}/(n)_j\) depends on \(\mathbb {E}Y^3\). Another consequence of formula (24) is that, with the help of (22), we can quickly compute the moments \(\mathbb {E}S_n^j\) for any \(n\ge j\) in terms of the corresponding moments for \(n<j\).

4 Lévy Processes and Centered Subordinators

Lévy processes are, in continuous time, the analogue to sums of independent identically distributed random variables in discrete time. It therefore seems plausible to obtain for such processes similar moment results to those given in the preceding section. Recall that a Lévy process \((Y(t))_{t\ge 0}\) is a stochastically continuous process starting at the origin and having independent stationary increments. A subordinator \((X(t))_{t\ge 0}\) is a Lévy process having right continuous nondecreasing paths.

Let \((W(t))_{t\ge 0}\) be a zero mean square integrable Lévy process whose characteristic function is given by (cf. Steutel and van Harn [17, p. 181])

$$\begin{aligned} \mathbb {E}e^{i\xi W(t)}=\exp \left( t\int _\mathbb {R} \left( e^{i\xi x}-1-i\xi x\right) \, K(dx)\right) ,\quad \xi \in \mathbb {R},\quad t\ge 0, \end{aligned}$$
(25)

where K(dx) is a Lévy measure on \(\mathbb {R}\), which puts mass 0 on \(\{0\}\) and satisfies

$$\begin{aligned} \int _\mathbb {R}|x|\, K(dx)<\infty ,\quad 0<\kappa ^2=\int _\mathbb {R}x^2\, K(dx)<\infty . \end{aligned}$$

The characteristic function (25) can be written as

$$\begin{aligned} \mathbb {E}e^{i\xi W(t)}=\exp \left( -\dfrac{t\kappa ^2 \xi ^2}{2}\mathbb {E}e^{i\xi \beta (2)U}\right) ,\quad \xi \in \mathbb {R},\quad t\ge 0, \end{aligned}$$
(26)

where \(\beta (2)\) is defined in (4) and U is a random variable independent of \(\beta (2)\), with distribution function

$$\begin{aligned} F_U(x)= \dfrac{1}{\kappa ^2} \int _{-\infty }^x y^2\, K(dy),\quad x\in \mathbb {R}. \end{aligned}$$

We see from (26) that \(\mathbb {E}W(t)=0\) and \(\mathbb {E}W^2(t)=t\kappa ^2\), \(t\ge 0\), so that \(\kappa ^2\) is the variance of W(1).

Now, let \((B(t))_{t\ge 0}\) be a standard Brownian motion on \(\mathbb {R}\), independent of \((W(t))_{t\ge 0}\) and define the Lévy process \((Y(t))_{t\ge 0}\) by setting

$$\begin{aligned} Y(t)=W(t)+\sigma B(t),\quad t\ge 0,\quad \sigma \ge 0. \end{aligned}$$
(27)

Observe that \(\mathbb {E}Y(t)=0\), \(\mathbb {E}Y^2(t)=t(\kappa ^2 +\sigma ^2)\), \(t\ge 0\), and

$$\begin{aligned} \mathbb {E}e^{i\xi Y(t)}=\exp \left( -\dfrac{t\sigma ^2 \xi ^2}{2}-\dfrac{t\kappa ^2 \xi ^2}{2} \mathbb {E}e^{i\xi \beta (2) U}\right) ,\quad \xi \in \mathbb {R},\quad t\ge 0, \end{aligned}$$
(28)

as follows from (26) and (27). Let V be a random variable uniformly distributed on [0, 1] and independent of \(\beta (2)\) and U. Then, we can rewrite (28) as

$$\begin{aligned} \begin{aligned} \mathbb {E}e^{i\xi Y(t)}&= \exp \left( -\dfrac{t(\sigma ^2+\kappa ^2)\xi ^2}{2}\, \mathbb {E}\exp \left( i\xi \beta (2) U 1_{\{V< \kappa ^2/(\sigma ^2+\kappa ^2)\}}\right) \right) \\&=\exp \left( -\dfrac{t(\sigma ^2+\kappa ^2)\xi ^2}{2} \, \mathbb {E}e^{i\xi \beta (2)T_\star }\right) ,\quad \xi \in \mathbb {R},\quad t\ge 0, \end{aligned} \end{aligned}$$
(29)

where

$$\begin{aligned} T_\star =U 1_{\{V< \kappa ^2/(\sigma ^2+\kappa ^2)\}}. \end{aligned}$$
(30)

On the other hand, a subordinator \((X(t))_{t\ge 0}\) is called centered if \(\mathbb {E}(X(t)-t)=0\) and \(\mathbb {E}(X(t)-t)^2<\infty , t\ge 0\). In such a case, the characteristic function of X(t) is then given by (cf. Steutel and van Harn [17, p. 107] and [2])

$$\begin{aligned} \mathbb {E}e^{i\xi X(t)}=\exp \left( t\mathbb {E}\frac{e^{i\xi T}-1}{T}\right) ,\quad \xi \in \mathbb {R},\quad t\ge 0, \end{aligned}$$

where T is a nonnegative random variable. Denote by \(\tau ^2=\mathbb {E}T\). This notation comes from the fact that \(\mathbb {E}(X(t)-t)^2=t\tau ^2, t\ge 0\). Consider the nonnegative random variable \(T^*\) whose distribution function is given by

$$\begin{aligned} F_{T^*} (y)=\dfrac{1}{\tau ^2} \int _0^y xF_T(dx), \quad y\ge 0, \end{aligned}$$

and equal to zero for \(y<0\), where \(F_T\) is the distribution function of T and it is assumed that \(\tau ^2>0\). In the case in which \(T=0\), a.s., we simply define \(T^*=0\), a.s. Observe that for any function f we have

$$\begin{aligned} \tau ^2\mathbb {E}f(T^*)=\mathbb {E}Tf(T). \end{aligned}$$

It turns out that (cf. [2])

$$\begin{aligned} \mathbb {E}e^{i\xi (X(t)-t)}=\exp \left( -\frac{t\tau ^2\xi ^2}{2}\mathbb {E}e^{i\xi \beta (2)T^*}\right) ,\quad \xi \in \mathbb {R},\quad t\ge 0, \end{aligned}$$
(31)

where the random variables \(T^*\) and \(\beta (2)\) are independent. The main difference between formulas (29) and (31) is that \(T_\star \) is real-valued, whereas \(T^*\) is nonnegative. We finally observe that if \((X(t))_{t\ge 0}\) is the standard Poisson process, then \(T=T^*=1\), whereas for the gamma process, the random variables T and \(T^*\) have the probability densities \(\rho (\theta )=e^{-\theta }\) and \(\rho ^*(\theta )=\theta e^{-\theta }, \theta \ge 0\), respectively.

Once representations (29) and (31) are given, we can obtain closed-form expressions for the moments of Y(t) and \(X(t)-t\) in a simple way, as the following result shows.

Theorem 4

Assume that \(T_\star \) and \(T^*\), appearing in (30) and (31), respectively, belong to \({\mathcal {G}}_0\). For any \(j\in \mathbb {N}_0\) and \(t\ge 0\), we have

$$\begin{aligned} g_j(t):=\frac{\mathbb {E}Y(t)^j}{t^{\lfloor j/2\rfloor }}=\sum _{m=0}^{\lfloor j/2\rfloor }\left( {\begin{array}{c}j\\ 2m\end{array}}\right) (\sigma ^2+\kappa ^2)^m\mathbb {E}Z^{2m} \mathbb {E}W_m(T_\star )^{j-2m}\frac{1}{t^{\lfloor j/2\rfloor -m}}, \end{aligned}$$

and

$$\begin{aligned} h_j(t):=\frac{\mathbb {E}(X(t)-t)^j}{t^{\lfloor j/2\rfloor }}=\sum _{m=0}^{\lfloor j/2\rfloor }\left( {\begin{array}{c}j\\ 2m\end{array}}\right) \mathbb {E}(\tau Z)^{2m}\mathbb {E}W_m(T^*)^{j-2m}\frac{1}{t^{\lfloor j/2\rfloor -m}}. \end{aligned}$$

Moreover, the functions \(g_{2j}(t)\) and \(h_j(t)\) are completely monotonic.

Proof

The identities in Theorem 4 follow by expanding the characteristic functions given in (29) and (31), and recalling (20). The last statements concerning complete monotonicity follow from the facts that \(\mathbb {E}W_m(T_\star )^{2(j-m)}\) and \(\mathbb {E}W_m(T^*)^{j-2m}\) are nonnegative for \(m\le \lfloor j/2\rfloor \). The proof is complete. \(\square \)

With respect to Theorem 4, similar comments to those made after Corollary 3 are valid. Details are omitted.

5 Edgeworth Expansions

Let \(y\in \mathbb {R}\) and let Z be a random variable having the standard normal density

$$\begin{aligned} g(y)=\frac{1}{\sqrt{2\pi }}e^{-y^2/2}. \end{aligned}$$

Denote by G(y) the standard normal distribution function. Recall that the Hermite polynomials \((H_n(y))_{n\ge 0}\) are defined by

$$\begin{aligned} g(y)H_n(y)=(-1)^n g^{(n)}(y). \end{aligned}$$

Since

$$\begin{aligned} \frac{1}{2\pi }\int _\mathbb {R}\mathbb {E}e^{i\zeta (Z-y)}d\xi =\frac{1}{2\pi }\int _\mathbb {R}e^{-i\xi y}e^{-\xi ^2/2}d\xi =\frac{1}{\sqrt{2\pi }}\mathbb {E}e^{-iyZ}=g(y), \end{aligned}$$

differentiation under the integral sign with respect to y gives us

$$\begin{aligned} \frac{1}{2\pi }\int _\mathbb {R}(i\xi )^n\mathbb {E}e^{i\zeta (Z-y)}d\xi =(-1)^n g^{(n)}(y)=g(y)H_n(y). \end{aligned}$$
(32)

Let \(Y\in {\mathcal {G}}_0\) be a real-valued random variable having an integrable characteristic function. Suppose that \(\mathbb {E}Y=0\) and \(\mathbb {E}Y^2=1\). Denote by \(F_n(y)\) the distribution function of \(S_n/\sqrt{n}\). Under such circumstances, it is well known (see, for instance, Petrov [15, p. 117]) that

$$\begin{aligned} F_n(y)-G(y)=-\frac{1}{2\pi }\int _\mathbb {R} e^{-i\xi y}\frac{\mathbb {E}e^{i\xi S_n/\sqrt{n}}-\mathbb {E}e^{i\xi Z}}{i\xi } d\xi . \end{aligned}$$
(33)

We will show that the Edgeworth expansion of \(F_n(y)-G(y)\) can be described in a simple way in terms of the Stirling numbers associated with the complex-valued random variable

$$\begin{aligned} {\widehat{Y}}=Y+iZ, \end{aligned}$$
(34)

where Y and Z are supposed to be independent. To this end, fix \(r=2,3,\ldots \) Consider the sets

$$\begin{aligned} \Delta =\{(m,j): 1\le m\le n,\quad j\ge m(r+1)\}, \end{aligned}$$

and

$$\begin{aligned} \Delta _k=\{(m,j)\in \Delta : j=2m+k\},\quad k=r-1,r,r+1,\ldots \end{aligned}$$
(35)

We are in a position to state the following.

Theorem 5

Let \(Y\in {\mathcal {G}}_0\) be a real-valued random variable having an integrable characteristic function. Assume that \(\mathbb {E}Y^k=\mathbb {E}Z^k, k=1,2,\ldots ,r\) for some \(r\ge 2\). For any \(n\in \mathbb {N}\) and \(y\in \mathbb {R}\), we have

$$\begin{aligned} F_n(y)-G(y)=-g(y)\sum _{k=r-1}^\infty \frac{1}{n^{k/2}}\sum _{(m,j)\in \Delta _k}\frac{S_{{\widehat{Y}}}(j,m)}{j!}\frac{(n)_m}{n^m}H_{j-1}(y). \end{aligned}$$
(36)

Proof

Let \(\xi \in \mathbb {R}\). By (34), the integrand in (33) can be written as

$$\begin{aligned} \mathbb {E}e^{i\xi (Z-y)}\frac{\left( \mathbb {E}e^{i\xi {\widehat{Y}}/\sqrt{n}}\right) ^n-1}{i\xi }. \end{aligned}$$
(37)

By Lemma 2, the random variable \({\widehat{Y}}\) belongs to \({\mathcal {G}}_r\). We therefore have from Theorem 2 and (35)

$$\begin{aligned} \begin{aligned} \frac{\left( \mathbb {E}e^{i\xi {\widehat{Y}}/\sqrt{n}}\right) ^n-1}{i\xi }&= \frac{1}{i\xi }\sum _{m=1}^n (n)_m \frac{\left( \mathbb {E}e^{i\xi {\widehat{Y}}/\sqrt{n}}-1\right) ^m}{m!}\\&= \frac{1}{i\xi }\sum _{m=1}^n (n)_m \sum _{j=m(r+1)}^\infty \frac{S_{{\widehat{Y}}}(j,m)}{j!}\left( \frac{i\xi }{\sqrt{n}}\right) ^j\\&=\sum _{k=r-1}^\infty \frac{1}{n^{k/2}}\sum _{(m,j)\in \Delta _k}\frac{S_{{\widehat{Y}}}(j,m)}{j!}\frac{(n)_m}{n^m}(i\xi )^{j-1}. \end{aligned} \end{aligned}$$

Hence, integrating (37) with respect to \(\xi \), the conclusion follows from (32). The proof is complete. \(\square \)

Fix \(r\ge 2\). Compared with the usual full Edgeworth expansions, Theorem 5 gives us an explicit and relatively simple expansion of \(F_n(y)-G(y)\), making clear, at the same time, that its order of magnitude is that of \(n^{-(r-1)/2}\), provided that the first moments of Y and Z (up to the order \(r\ge 2\)) coincide. The coefficients in this expansion depend on the Stirling numbers \(S_{{\widehat{Y}}}(j,m)\) associated with the complex-valued random variable \({\widehat{Y}}\) defined in (34). As noted after Theorem 1, such numbers are actually real numbers, which can be evaluated by means of Theorem 2.

For instance, let us evaluate the leading coefficient in (36). As follows from (35), we have \(\Delta _{r-1}=\{(1,r+1)\}\). Thus, the leading coefficient in (36) is equal to

$$\begin{aligned} -g(y)H_r(y)\frac{S_{{\widehat{Y}}}(r+1,1)}{(r+1)!}=-g(y)H_r(y)\frac{\mathbb {E}{\widehat{Y}}^{r+1}}{(r+1)!}, \end{aligned}$$

as follows from Theorem 2. On the other hand, if \(r+1\) is odd, it can be checked from (34) that \(\mathbb {E}{\widehat{Y}}^{r+1}=\mathbb {E}Y^{r+1}\), whereas if \(r+1=2s\), we have from (20) and the moment assumptions in Theorem 5

$$\begin{aligned} \begin{aligned} \mathbb {E}{\widehat{Y}}^{r+1}&=\mathbb {E}{\widehat{Y}}^{2s}=\sum _{k=0}^{2s}\left( {\begin{array}{c}2s\\ k\end{array}}\right) \mathbb {E}Y^k\mathbb {E}(iZ)^{2s-k}\\&= \mathbb {E}Y^{2s}+\sum _{l=0}^{s-1}\left( {\begin{array}{c}2s\\ 2l\end{array}}\right) \mathbb {E}Z^{2l}(-1)^{s-l}\mathbb {E}Z^{2(s-l)}\\&= \mathbb {E}Y^{2s}+\frac{(2s)!}{s!2^s}\sum _{l=0}^{s-1}\left( {\begin{array}{c}s\\ l\end{array}}\right) (-1)^{s-l}=\mathbb {E}Y^{2s}-(-1)^s\mathbb {E}Z^{2s}. \end{aligned} \end{aligned}$$

6 Cumulants

Recall that the cumulant generating function of a random variable \(Y\in {\mathcal {G}}_0\) is defined as

$$\begin{aligned} K_Y(z)=\log \mathbb {E}e^{zY}=\sum _{j=1}^\infty \kappa _j(Y)\frac{z^j}{j!}, \end{aligned}$$
(38)

where the coefficients \((\kappa _j(Y))_{j\ge 1}\) are called the cumulants of Y. Such cumulants can be written in terms of the Stirling numbers \(S_Y(j,m)\), as shown in the following result.

Theorem 6

Let \(Y\in {\mathcal {G}}_0\). For any \(j\in \mathbb {N}\), we have

$$\begin{aligned} \kappa _j(Y)=\sum _{m=1}^j (-1)^{m-1}(m-1)!S_Y(j,m)=\sum _{k=1}^j \left( {\begin{array}{c}j\\ k\end{array}}\right) \frac{(-1)^{k-1}}{k}\mathbb {E}S_k^j. \end{aligned}$$
(39)

Proof

Using the expression

$$\begin{aligned} \log (1+x)=\sum _{m=1}^\infty (-1)^{m-1}\frac{x^m}{m},\quad |x|<1, \end{aligned}$$

and choosing z in a neighborhood of the origin so that \(|\mathbb {E}e^{zY}-1|<1\), we get

$$\begin{aligned} \begin{aligned} K_Y(z)&=\log \left( 1+\mathbb {E}e^{zY}-1\right) =\sum _{m=1}^\infty (-1)^{m-1}(m-1)!\frac{\left( \mathbb {E}e^{zY}-1\right) ^m}{m!}\\&=\sum _{m=1}^\infty (-1)^{m-1}(m-1)!\sum _{j=m}^\infty S_Y(j,m)\frac{z^j}{j!}\\&=\sum _{j=1}^\infty \frac{z^j}{j!}\sum _{m=1}^j(-1)^{m-1}(m-1)!S_Y(j,m). \end{aligned} \end{aligned}$$

In view of (38), this shows the first equality in (39). The second one readily follows from definition (3) and the well-known combinatorial identity

$$\begin{aligned} \sum _{l=0}^p \left( {\begin{array}{c}s+l\\ l\end{array}}\right) =\left( {\begin{array}{c}s+p+1\\ p\end{array}}\right) ,\quad p,s\in \mathbb {N}_0. \end{aligned}$$

The proof is complete. \(\square \)

We finally mention that, in certain particular cases, we find for the cumulants simpler formulas than those given in (39). For instance, in the case of the Lévy processes considered in (29), it can be checked that

$$\begin{aligned} \kappa _j(Y(t))=t(\sigma ^2+\kappa ^2)\mathbb {E}T_\star ^{j-2},\quad j=2,3,\ldots , t\ge 0. \end{aligned}$$

A similar formula holds for the centered subordinators defined in (31).