Abstract
We explore two aspects of geometric approximation via a coupling approach to Stein’s method. Firstly, we refine precision and increase scope for applications by convoluting the approximating geometric distribution with a simple translation selected based on the problem at hand. Secondly, we give applications to several stochastic processes, including the approximation of Poisson processes with random time horizons and Markov chain hitting times. Particular attention is given to geometric approximation of random sums, for which explicit bounds are established. These are applied to give simple approximations, including error bounds, for the infinite-horizon ruin probability in the compound binomial risk process.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction and Motivation
Approximation by the geometric distribution and its generalisations have continued to receive particular attention in recent years, especially in conjunction with Stein’s method, a powerful technique for probability approximation that has been used successfully for a variety of theoretical and applied topics. Stein’s method was first introduced in Stein (1972); see also Ross (2011) for a more recent introduction to the area. Stein’s method for geometric approximation was developed by Barbour and Grübel (1995) and Peköz (1996). We refer the reader to Peköz et al. (2013) and references therein for more recent developments.
Our goal in this paper is twofold. Our first aim is to consider how we can improve upon geometric approximation by incorporating a convolution with another random variable. Though a geometric approximation can yield effective error bounds in some applications, extra precision can be gained, and the range of useful applications increased, by using an approximating distribution that incorporates some additional parameters. Examples of this approach include negative binomial approximation (see Ross (2013); Yu et al. (2023) and references therein) and approximation by a geometric sum (Daly, 2010, 2016, 2019). Such approaches naturally lead to additional technical complexity in the evaluation of the approximating distribution or the corresponding error bounds. Our approach of convoluting the approximating geometric distribution with an independent translation is an alternative to these ideas. The mass function of this convolution must be kept simple to use, otherwise the benefits are lost. We are motivated here by the work of Röllin (2007) on translated Poisson approximation, where better error bounds are obtained than using a traditional Poisson approximation. Our translation will allow us to introduce additional parameters in the approximating distribution which can then be chosen to gain precision, at the cost of some additional complexity in the approximation. Unlike the usual approach to translated Poisson approximation where the translation is by a constant, we will here allow translation by an independent random variable.
Our second goal is to consider geometric approximation of random sums (i.e., of sums of a random number of random variables) and related applications to risk processes. The case in which the number of terms in the sum is itself geometrically distributed has previously been treated by Peköz et al. (2013); here we consider the more general case. Our results are applied to give a simple approximation, with explicit error bounds, of the infinite-horizon ruin probability in the compound binomial risk process (Gerber, 1988). As further applications of our general results, we will also give explicit error bounds in the approximation of the number of points in a Poisson process over a random time horizon, and of Markov chain hitting times.
In these applications, and in particular in the approximation of random sums, there are a number of other approximating distributions available, including compound Poisson approximations (see, for example, Vellaisamy and Chaudhuri (1996)) which offer a flexible approximation framework. A geometric distribution is a simple special case of a compound Poisson distribution. This simplicity leads to more simply stated and easily calculated error bounds compared to a more general compound Poisson setting. As we shall see in the examples and applications of Sections 2 and 3 below, there are distributions of interest which are close to geometric, so this approximation is an appropriate and useful one to consider here.
Recall that \(Y\sim \text {Geom}(p)\) has a geometric distribution if \(\mathbb {P}(Y=k)=p(1-p)^k\) for \(k\in \mathbb {Z}^+=\{0,1,\ldots \}\). In evaluating the approximation error in the work that follows, we will employ the total variation distance between two non-negative, integer-valued random variables W and Y, which can be defined in any of the following equivalent ways:
where \(\mathcal {L}(W)\) denotes the law of W, and the infimum is taken over all couplings of W and Y with the required marginals.
The remainder of this paper is organised as follows. In Section 2 we introduce Stein’s method for geometric approximation, and use these techniques to derive results in the setting where we approximate by the convolution of a geometric distribution and an independent translation. This is illustrated by applications to Poisson processes and Markov chain hitting times. We then consider geometric approximation of random sums in Section 3, and apply these results to derive simple approximations for the ruin probability in the compound binomial risk process.
2 Translated Geometric Approximation
Following Stein’s method for geometric approximation (Barbour and Grübel, 1995; Peköz, 1996), for each \(A\subseteq \mathbb {Z}^+\) we let \(g_A:\mathbb {Z}^+\rightarrow \mathbb {R}\) be such that \(g_A(k)=0\) for \(k\le 0\) and
for some \(p\in (0,1]\), where \(Y\sim \text{ Geom }(p)\). Recall that Stein’s method proceeds by replacing k in Eq. 2 by the random variable W to be approximated, taking expectations, and bounding the expression thus obtained from the right-hand side of Eq. 2. Essential ingredients in the final part of this procedure include bounds on the behaviour of the functions \(g_A\). For example,
by Lemma 4.2 of Daly (2016) and Lemma 1 of Peköz (1996).
In the work that follows, we let W denote the non-negative, integer-valued random variable we wish to approximate. We consider the approximation of W by \(Y+T\), where \(Y\sim \text{ Geom }(p)\) for some p, and T is our random translation independent of W and Y. Our results will employ a coupling construction based on that used by Peköz (1996) and Daly (2010). To that end, we define the random variable V to be such that
emphasising that V will depend on both T and W.
With this setup, we may obtain the following
Theorem 1
Let W and T be independent integer-valued random variables with \(W\ge 0\). Suppose V is as defined by Eq. 4 and let \(Y\sim \text{ Geom }(p)\) be independent of T, where \(p=\mathbb {P}(W\le T)\). Then
Proof
Since T is independent of W and Y,
Motivated by the argument of (3.11) of Röllin (2007), we now write
where we recall that \(g_A(k)=0\) for \(k\le 0\) and use Eq. 2 with \(k=W-T\) for the final expression. Now, conditioning on the value of W,
using the definitions of \(g_A\) and V. Our bound Eq. 5 then follows from Eq. 3 when writing
Similarly, for Eq. 6 we let \(I(\cdot )\) denote an indicator function and write
using the definition Eq. 1 of total variation distance and the bound Eq. 3. \(\square \)
Remark 1
Depending on the situation, it is sometimes more natural to define a geometric random variable on \(\{1,2\ldots \}\) rather than \(\{0,1,\ldots \}\). In our framework it is straightforward to switch between the two definitions by simply altering the definition of T.
In the remainder of this section we consider two applications of Theorem 1, to approximation of the number of points observed in a Poisson process over a random time horizon and of Markov chain hitting times. For later use, we recall that a random variable U is said to be stochastically larger than a random variable V if \(\mathbb {P}(U>t)\ge \mathbb {P}(V>t)\) for all t, and that this is equivalent to the existence of a probability space on which we may construct the pair \((U^\prime ,V^\prime )\) such that \(\mathcal {L}(U^\prime )=\mathcal {L}(U)\), \(\mathcal {L}(V^\prime )=\mathcal {L}(V)\), and \(U^\prime \ge V^\prime \) almost surely; see Theorem 1.A.1 of Shaked and Shanthikumar (2007). We denote this by \(U\ge _{st}V\).
2.1 Application: Poisson Processes with Random Time Horizon
Let \(\{N(t):t\ge 0\}\) be a homogeneous Poisson process of rate \(\lambda \), and \(\tau \) be a non-negative random variable independent of this process. We consider the approximation of \(W=N(\tau )\) by \(Y+T\), where Y has a geometric distribution and T is a non-negative, integer-valued random variable independent of W and Y. In this we are motivated by Daly (2016), who previously considered geometric-type approximations in this setting, and by applications considered by Asmussen et al. (2002) and others in which Poisson processes with random time horizons play a role. We expect a geometric-type approximation to be relevant in cases of practical importance, since if \(\tau \) has an exponential distribution then \(N(\tau )\) is precisely geometrically distributed.
Let \(\xi _0=0\) and, for \(i\ge 1\), let \(\xi _i\) be the time of the ith event in our Poisson process, so that \(N(t)=\max \{i:\xi _i\le t\}\). We define \(V=N(\sigma _T)\), where
and \(\sigma _T\) depends on the random variables \(\{N(t):t\ge 0\}\), through \(\xi _{T+1}\), and \(\tau \). This V satisfies Eq. 4. To see this, note that conditioning on \(N(\tau )>T\) is equivalent to conditioning on the \((T+1)\)th event occurring before time \(\tau \), i.e., that \(\xi _{T+1}\le \tau \). Conditioning on this, the number of events we have seen up to time \(\tau \) is then \(T+1+V\).
We thus obtain from Eq. 5 that
where \(Y\sim \text{ Geom }(p)\) and \(p=\mathbb {P}(\tau <\xi _{T+1})\).
Consider, for simplicity, the case where \(T=0\) almost surely, and write \(\sigma =\sigma _0\). Note that
and
where \(p=\mathbb {P}(\tau <\xi _1)=\mathbb {E}[e^{-\lambda \tau }]\).
In the case where \(\sigma \ge _{st}\tau \) the bound Eq. 7 simplifies considerably, since we then have
Similarly, if \(\tau \ge _{st}\sigma \) we have
Observe that if \(\tau \) satisfies the new worse than used (NWU) property that \(\mathbb {P}(\tau>u)\mathbb {P}(\tau>v)\le \mathbb {P}(\tau >u+v)\) for all \(u,v\ge 0\), the independence of \(\tau \) and \(\xi _1\) combined with Eq. 8 implies that \(\sigma \ge _{st}\tau \). Similarly, if \(\tau \) satisfies the new better than used (NBU) property that \(\mathbb {P}(\tau>u)\mathbb {P}(\tau>v)\ge \mathbb {P}(\tau >u+v)\) for all \(u,v\ge 0\), we have that \(\tau \ge _{st}\sigma \). By combining the bound Eq. 7 with the choice \(p=\mathbb {E}[e^{-\lambda \tau }]\) and the expectations calculated in Eqs. 9 and 10 we thus have the following
Theorem 2
Let \(\{N(t):t\ge 0\}\) be a homogeneous Poisson process of rate \(\lambda \). Let \(\tau \) be a non-negative random variable independent of this process.
-
(a)
If \(\tau \) is NBU,
$$\begin{aligned} d_{TV}(\mathcal {L}(N(\tau )),\text {Geom}(\mathbb {E}[e^{-\lambda \tau }]))\le 1-\mathbb {E}[e^{-\lambda \tau }](1+\lambda \mathbb {E}[\tau ])\,. \end{aligned}$$ -
(b)
If \(\tau \) is NWU,
$$\begin{aligned} d_{TV}(\mathcal {L}(N(\tau )),\text {Geom}(\mathbb {E}[e^{-\lambda \tau }]))\le \mathbb {E}[e^{-\lambda \tau }](1+\lambda \mathbb {E}[\tau ])-1\,. \end{aligned}$$
If \(\tau \) is exponentially distributed (which is trivially both NBU and NWU) then \(\tau \) and \(\sigma =\sigma _0\) as above have the same distribution, and the upper bounds of Theorem 2 are each zero, reflecting the fact that \(N(\tau )\) has a geometric distribution.
The upper bound of Theorem 2(a) was previously established in Section 2.2 of Daly (2016), but only under the stronger condition that \(\tau \) has increasing failure rate.
We conclude this section with two examples illustrating Theorem 2.
Example 1
Let \(\tau \) have a gamma distribution, and note that \(N(\tau )\) is thus negative binomial. For simplicity of presentation we normalise \(\tau \) to have unit mean, with density function \(x\mapsto \frac{\beta ^\beta }{\Gamma (\beta )}x^{\beta -1}e^{-\beta x}\) for \(x>0\), where \(\beta >0\) is a parameter and \(\Gamma (\cdot )\) is the usual gamma function. We have that \(\mathbb {E}[e^{-\lambda \tau }]=\left( 1+\frac{\lambda }{\beta }\right) ^{-\beta }\) in this case. Note that \(\tau \) has decreasing failure rate when \(\beta <1\) (and is thus NWU), constant failure rate when \(\beta =1\) (the exponential case), and increasing failure rate when \(\beta >1\) (and is thus NBU). Applying both parts of Theorem 2 we have
The upper bound Eq. 11 is illustrated in Fig. 1 for several values of \(\lambda \). As expected, this upper bound is zero when \(\beta =1\), and is small when \(\beta \) is close to 1.
Example 2
Let \(\tau \) be uniformly distributed on the interval [a, b] for some \(0\le a<b\), so that \(\tau \) has increasing failure rate (and is thus NBU) and satisfies \(\mathbb {E}[e^{-\lambda \tau }]=\frac{e^{-\lambda a}-e^{-\lambda b}}{\lambda (b-a)}\) for all \(\lambda >0\). Hence, by Theorem 2(a),
If \(\lambda a\) and \(\lambda b\) are small, writing \(e^{-\lambda a}\approx 1-\lambda a\) and \(e^{-\lambda b}\approx 1-\lambda b\) we see that this upper bound is small when \(a+b\approx \frac{2(1-\lambda ^2)}{\lambda }\).
2.2 Application: Markov Chain Hitting Times
We now let W be the hitting time of an ergodic Markov chain to a set of states A. That is, we let \((\xi _0,\xi _1,\ldots )\) be a Markov chain and \(W=\min \{i\ge 0:\xi _i\in A\}\) for some fixed subset A of the state space. We will write \(P_{ij}^{(n)}=\mathbb {P}(\xi _n=j|\xi _0=i)\) for the n-step transition probabilities and \(\pi \) for the stationary distribution of our Markov chain, with \(\pi (A)\) denoting the measure of a set A according to \(\pi \); abusing notation, we will write \(\pi _i\) for \(\pi (\{i\})\).
For some distribution F on this state space, we let \(\xi _0\sim F\), and define \(T=T(F)\) to be a stationary time for this initial distribution, independent of W. That is, T is such that \(\mathcal {L}(\xi _T|\xi _0\sim F)\) is \(\pi \). In line with Theorem 1, we consider the approximation of W by \(Y+T\), where \(Y\sim \text{ Geom }(p)\) is a geometric random variable with parameter \(p=\mathbb {P}(W\le T)\) which is independent of T. We will use Eq. 6 to give a bound. To that end, we construct the random variables \(W-T\) and V as follows: Consider the Markov chain \((X,\xi _0^\prime ,\xi _1^\prime ,\ldots )\), where the initial state X is sampled according to \(\pi \). Since T is a stationary time, we may construct \(W-T\) as the number of time steps required to hit A from an initial stationary distribution. That is, we write \(W-T=\min \{i\ge 0:\xi _i^\prime \in A\}\). Now consider another Markov chain \((Y,\xi _0^{\prime \prime },\xi _1^{\prime \prime },\ldots )\) whose initial state Y is sampled according to \(\pi \) restricted to \(A^\text {c}\), the complement of A. We may then write \(V=\min \{i\ge 0:\xi _i^{\prime \prime }\in A\}\) since, in conditioning on the event that \(W>T\), we consider the number of steps from such an initial state until we hit A in addition to the \(T+1\) steps already taken to reach the state \(\xi _0^{\prime \prime }\).
With these random variables, the maximal coupling argument of Theorem 2 of Peköz (1996) gives
The following bound is then immediate from Eq. 6.
Theorem 3
Let \((\xi _0,\xi _1,\ldots )\) be the Markov chain above, \(W=\min \{i\ge 0:\xi _i\in A\}\), and T be a stationary time as above. With \(Y\sim \text{ Geom }(p)\) for \(p=\mathbb {P}(W\le T)\),
This result generalises Theorem 2 of Peköz (1996), where the hitting time from an initial stationary distribution only is considered, in which case our T is equal to zero almost surely and \(p=\pi (A)\).
3 Geometric Approximation of Random Sums and the Compound Binomial Risk Process
Throughout this section we let \(W=\sum _{i=1}^NX_i\), where \(X_1,X_2,\ldots \) are independent and identically distributed, positive integer-valued random variables with \(\mathbb {E}[X_1]=\mu \ge 1\). We let N be a non-negative, integer-valued random variable independent of the \(X_i\). For simplicity we restrict ourselves to the case where the \(X_i\) are identically distributed, but this condition can easily be relaxed.
We will consider the approximation of W by a geometric random variable Y in Section 3.1, and then apply our results in Section 3.2 to derive simple and explicit bounds for the infinite-horizon ruin probability in the compound binomial risk process.
In this section we focus on geometric approximation (i.e., \(T=0\) almost surely in the setting of Theorem 1), since construction of the random variable V defined in Eq. 4 seems to be natural here only when \(T=0\). This case is also sufficient to give simple and explicit bounds on our ruin probability of interest, which is the principal application of this section.
3.1 Geometric Approximation of Random Sums
With W the random sum defined above, we now derive geometric approximation bounds for W. For later use, we define \(N^{(0)}\) such that \(\mathcal {L}(N^{(0)})=\mathcal {L}(N|N>0)\).
We consider the approximation of W by \(Y\sim \text{ Geom }(p)\). In line with the discussion in Section 2 in the case where \(T=0\) almost surely, we choose the parameter p to be
and V to be such that
so that we may choose \(V=\sum _{j=1}^{N^{(0)}}X_j-1\). With this choice we have that
since \(N^{(0)}\ge _{st}N\), and so we may construct these on the same probability space in such a way that \(N^{(0)}\ge N\) almost surely.
Furthermore, if \(N^{(0)}\ge _{st}N+1\) then we may construct our random variables such that \(\sum _{i=N+1}^{N^{(0)}}X_i\) is almost surely greater than 1, since each \(X_i\) is at least 1 almost surely. In this case we thus have that
From the definition, we note that \(N^{(0)}\ge _{st}N+1\) if
for all \(j\in \mathbb {Z}^+\). That is, \(N^{(0)}\ge _{st}N+1\) if N is larger than Y in the hazard rate ordering; see Section 1.B of Shaked and Shanthikumar (2007).
Combining all these ingredients, we may apply Eq. 5 with \(T=0\) to give
Theorem 4
Let \(W=\sum _{i=1}^NX_i\) be as above and p be given by Eq. 12. Then
Moreover, if Eq. 13 holds for all \(j\in \mathbb {Z}^+\), then
One case which has received particular attention in the literature is that in which \(N\sim \text{ Geom }(r)\) has a geometric distribution, so that W is a geometric sum. Note that in this case \(\mathcal {L}(N^{(0)})=\mathcal {L}(N+1)\), so that the bound Eq. 14 applies. Geometric approximation for geometric sums has also been considered by Peköz et al. (2013), who show in their Theorem 3.2 that
where \(p^\prime =\frac{r}{r+\mu (1-r)}\) and u is such that \(d_{TV}(\mathcal {L}(X_1),\mathcal {L}(X_1+1))\le 1-u\). Note that this parameter \(p^\prime \) is chosen such that the mean of the approximating geometric distribution matches that of W; on the other hand our choice is made such that each have the same probability of taking the value zero. In the setting where we have no information on the smoothness of the \(X_i\) in the form of a bound on u, our geometric approximation bound improves upon this for sufficiently large \(\mathbb {E}[X_1^2]\), and also applies in the case where N is not geometric. We note also that both geometric approximation results have the desirable property that the upper bound is zero whenever \(\mu =1\). Finally, we refer to Peköz and Röllin (2011) for related results on the exponential approximation of random sums in the case where the \(X_i\) are no longer assumed to be integer-valued.
3.2 Application: Compound Binomial Risk Process
We now consider how our Theorem 4 may be applied to the compound binomial risk process in discrete time (Gerber, 1988), defined as follows: Let \(m\in \mathbb {Z}^+\) and let \(\eta ,\eta _1,\eta _2,\ldots \) be independent and identically distributed random variables supported on \(\mathbb {Z}^+\). The risk process \(\{U_t:t=0,1,\ldots \}\) is defined by
where we write \(q=\mathbb {E}[\eta ]\) and make the assumption that \(q<1\) to ensure a positive drift of this process. We will also assume the existence of second and third moments of \(\eta \) as required. Letting \(\tau =\min \{t\ge 1:U_t\le 0\}\), our interest is in the infinite-horizon ruin probability \(\psi (m)\) defined by
This ruin probability is given by a discrete Pollaczek–Khinchine formula: We have \(\psi (0)=q\) and
for \(m=1,2,\ldots \), where \(M\sim \text{ Geom }(1-q)\) and \(Y,Y_1,Y_2,\ldots \) are independent and identically distributed with \(\mathbb {P}(Y=j)=q^{-1}\mathbb {P}(\eta >j)\) for \(j=0,1,\ldots \); see Proposition 2.6 of Santana and Rincón (2020). We note that Santana and Rincón (2020) gives some numerical approximations for \(\psi (m)\), but without any explicit error bounds. Our goal in this section is to give explicit approximations for \(\psi (m)\), complete with error bounds, that are significantly easier to work with than the tail probabilities in Eq. 17.
The geometric sum in Eq. 17 does not quite fit into the framework of Theorem 4, since here the summands are allowed to take the value zero with positive probability. We therefore rewrite this by letting \(X,X_1,X_2\ldots \) be independent and identically distributed with
for \(j=1,2,\ldots \), and letting \(N\sim \text{ Geom }(r)\) with
The following lemma is well known (it follows, for example, from a stronger observation on page 165 of Michel (1987)), but we include its short proof to keep our exposition self-contained.
Lemma 1
Letting \(W=\sum _{i=1}^NX_i\), with N and the \(X_i\) as above, we have \(\mathcal {L}(W)=\mathcal {L}\left( \sum _{i=1}^MY_i\right) \).
Proof
Let \(K_1,K_2,\ldots \) be independent and identically distributed, with \(K_j\sim \text{ Bin }(j,\mathbb {P}(Y>0))\) having a binomial distribution for each \(j=1,2,\ldots \). For \(m\ge 0\) we write
Of the j summands within the latter probability, we have \(K_j\) that are non-zero, with the remaining \(j-K_j\) taking the value zero. Conditioning on these \(K_j\) and noting that those \(Y_i\) that are conditioned to be non-zero have the same distribution as X, we obtain the following, where we write \(a=\mathbb {P}(Y=0)\):
as required. \(\square \)
In order to state a bound on \(\psi (m)\) we will need some further properties of the random variable X, which we collect in the following lemma.
Lemma 2
With X as above,
and
Proof
For \(l=1,2\) we write
from which the first two expressions in the lemma follow. For the total variation distance, the definition Eq. 1 gives
as required. \(\square \)
We are now in a position to state the main result of this section.
Theorem 5
With \(\psi (m)\) the ruin probability defined in Eq. 16 above,
and
for \(m=1,2,\ldots \), where
Proof
From Lemma 1 and the Pollaczeck–Khinchine formula Eq. 17 we may write that \(\psi (m)=\mathbb {P}(W\ge m)\) for \(m=1,2,\ldots \). With this representation, we employ the bound Eq. 14 of Theorem 4 in conjunction with the formula for \(\mathbb {E}[X]\) in Lemma 2 to obtain that
The proof of Eq. 19 is completed by noting that W is stochastically larger than our approximating geometric random variable, and so \(\mathbb {P}(W\ge m)\ge \left( \frac{q-\mathbb {P}(\eta >0)}{\mathbb {P}(\eta =0)}\right) ^m\).
For Eq. 20 we will use Eq. 15. In line with the notation there, and using the total variation distance given in Lemma 2, we write \(1-u=\frac{\mathbb {P}(\eta>0)}{2(q-\mathbb {P}(\eta >0))}\), so that
and so, with r given by Eq. 18,
The result then follows from Eq. 15 and Lemma 2. \(\square \)
Note that, since they are based on geometric approximation of a geometric sum, we would expect the bounds in Theorem 5 to perform well when the summands X are typically not much larger than one, that is, when \(\mathbb {P}(\eta >2)\) is small. To conclude this section, we illustrate this point and the bounds of Theorem 5 with three examples.
Example 3
Suppose that \(\eta \) is supported on \(\{0,1,2\}\), with \(\mathbb {P}(\eta =0)=\alpha _0\), \(\mathbb {P}(\eta =2)=\alpha _2\), and \(\mathbb {P}(\eta =1)=1-\alpha _0-\alpha _2\). We will assume that \(\alpha _2<\alpha _0\) to ensure that \(q<1\). In this case, both the bounds of Theorem 5 identify that \(\psi (m)=(\alpha _2/\alpha _0)^m\). We note, though, that Eq. 19 applies for all \(\alpha _2\in [0,\alpha _0)\), while the upper bound of Eq. 20 requires \(\alpha _2>0\) to be well-defined. In this example the expression for \(\psi (m)\) also follows more directly from the observation that \((U_0,U_1,\ldots )\) is a random walk on the integers with initial value \(U_0=m\), and independent increments supported on \(\{-1,0,1\}\) and distributed as \(1-\eta \).
Example 4
In the setting where \(\eta \sim \text{ Geom }(\alpha )\) has a geometric distribution, with \(\alpha >1/2\) to ensure that \(q<1\), it is known that
See Section 2 of Santana and Rincón (2020). Neither of the estimates in Theorem 5 yield exactly this value; the approximations given by Eqs. 19 and 20 are
respectively. The first of these gives the correct value for \(m=1\), but the second is a better approximation for larger m.
Example 5
Suppose \(\eta \) has a mixed Poisson distribution. That is, given some non-negative random variable \(\xi \), suppose that \(\eta |\xi \sim \text{ Pois }(\xi )\) has a Poisson distribution. We will illustrate our bound Eq. 19 in this case; the bound Eq. 20 may be evaluated similarly, but we omit that to keep the discussion concise. We have that \(q=\mathbb {E}[\xi ]\), which we assume is less than 1. We also have \(\mathbb {P}(\eta =0)=\mathbb {E}[e^{-\xi }]\), and \(\mathbb {E}[\eta (\eta -1)]=\mathbb {E}[\xi ^2]\). Hence, the approximation error in Eq. 19, i.e., the final term on the right-hand side, is
For example, if \(\eta \sim \text{ Pois }(\lambda )\) has a Poisson distribution for some small \(\lambda >0\), writing \(e^{\pm \lambda }\approx 1\pm \lambda \), this error is approximately \(\frac{1}{2}\lambda ^2(1+\lambda )\). The approximation error Eq. 21 in this case is illustrated in Figure 2.
As a second example, suppose instead that \(\xi \) has a gamma distribution with density \(x\mapsto \frac{\beta ^\alpha }{\Gamma (\alpha )}x^{\alpha -1}e^{-\beta x}\) for \(x>0\), where \(\alpha >0\) and \(\beta >0\) are parameters, so that \(\eta \) has a negative binomial distribution. Then our approximation error is
This converges to zero as \(\alpha \rightarrow 0\) for fixed \(\beta \), or as \(\beta \rightarrow \infty \) for fixed \(\alpha \), for example. This can be seen on Fig. 3, which shows the approximation error Eq. 22 as a function of \(\beta \) for various values of \(\alpha \).
Data Availability
No datasets were generated or analysed during the current study.
References
Stein C (1972) A bound for the error in the normal approximation to the distribution of a sum of dependent random variables. In: Proc. Sixth Berkeley Symp. Math. Statis. Prob., vol 2, pp 583–602. Univ. California Press
Ross N (2011) Fundamentals of Stein’s method. Probab Surv 8:210–293. https://doi.org/10.1214/11-PS182
Barbour AD, Grübel R (1995) The first divisible sum. J Theoret Probab 8:39–47. https://doi.org/10.1007/BF02213453
Peköz EA (1996) Stein’s method for geometric approximation. J Appl Probab 33:707–713. https://doi.org/10.2307/3215352
Peköz EA, Röllin AR, Ross N (2013) Total variation error bounds for geometric approximation. Bernoulli 19:610–632. https://doi.org/10.3150/11-BEJ406
Ross N (2013) Power laws in preferential attachment graphs and Stein’s method for the negative binomial distribution. Adv Appl Probab 45:876–893. https://doi.org/10.1239/aap/1377868543
Yu L, Daly F (2023) Johnson O (2023) A negative binomial approximation in group testing. Probab Engrg Inform Sci 37:973–996. https://doi.org/10.1017/S026996482200033X
Daly F (2010) Stein’s method for compound geometric approximation. J Appl Probab 47:146–156. https://doi.org/10.1239/jap/1269610822
Daly F (2016) Compound geometric approximation under a failure rate constraint. J Appl Probab 53:700–714. https://doi.org/10.1017/jpr.2016.35
Daly F (2019) On strong stationary times and approximation of Markov chain hitting times by geometric sums. Statist Probab Lett 150:74–80. https://doi.org/10.1016/j.spl.2019.02.017
Röllin AR (2007) Translated Poisson approximation using exchangeable pair couplings. Ann Appl Probab 17:1596–1614. https://doi.org/10.1214/105051607000000258
Gerber HU (1988) Mathematical fun with the compound binomial process. Astin Bull 18:161–168. https://doi.org/10.2143/AST.18.2.2014949
Vellaisamy P, Chaudhuri B (1996) Poisson and compound Poisson approximations for random sums of random variables. J Appl Probab 33:127–137. https://doi.org/10.2307/3215270
Shaked M, Shanthikumar JG (2007) Stochastic Orders. Springer, ork
Asmussen S, Avram F, Usabel M (2002) Erlangian approximations for finite-horizon ruin probabilities. Astin Bull 32:267–281. https://doi.org/10.2143/AST.32.2.1029
Peköz EA, Röllin A (2011) New rates for exponential approximation and the theorems of Rényi and Yaglom. Ann Probab 39:587–608. https://doi.org/10.1214/10-AOP559
Santana DJ, Rincón L (2020) Approximations of the ruin probability in a discrete time risk model. Mod Stoch Theory Appl 7:221–243. https://doi.org/10.15559/20-VMSTA158
Michel R (1987) An improved error bound for the compound Poisson approximation of a nearly homogeneous portfolio. Astin Bull 17:165–169. https://doi.org/10.2143/AST.17.2.2014971
Acknowledgements
The authors thank the Editor and two anonymous reviewers for their careful readings of an earlier version of the paper and suggestions which improved the presentation of the work.
Funding
F. Daly thanks the Belgian Fonds de la Recherche Scientifique for financial support for a visit to the Université Libre de Bruxelles where this work started. C. Lefèvre has received support from the DIALog Research Chair under the aegis of the Risk Foundation, an initiative of CNP Assurances.
Author information
Authors and Affiliations
Contributions
Both authors (F. Daly and C. Lefèvre) contributed to the conception, design and preparation of this work. Research was carried out jointly by both authors. Both authors contributed to the preparation of the manuscript, and have read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing of interest
C. Lefèvre serves on the editorial board of Methodology and Computing in Applied Probability.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Daly, F., Lefèvre, C. On Geometric-type Approximations with Applications. Methodol Comput Appl Probab 27, 4 (2025). https://doi.org/10.1007/s11009-024-10130-w
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11009-024-10130-w
Keywords
- Stein’s method
- Poisson process with random time horizon
- Markov chain hitting time
- Compound binomial risk process