This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

REFINED DISTRIBUTIONAL LIMIT THEOREMS
FOR COMPOUND SUMS

Vsevolod K. Malinovskii Central Economics and Mathematics Institute (CEMI) of Russian Academy of Science, 117418, Nakhimovskiy prosp., 47, Moscow, Russia admin@actlab.ru http://www.actlab.ru
Abstract.

The paper is a sketch of systematic presentation of distributional limit theorems and their refinements for compound sums. When analyzing, e.g., ergodic semi–Markov systems with discrete or continuous time, this allows us to separate those aspects that lie within the theory of random processes from those that relate to the classical summation theory. All these limit theorems are united by a common approach to their proof, based on the total probability rule, auxiliary multidimensional limit theorems for sums of independent random vectors, and (optionally) modular analysis.

Key words and phrases:
Compound sums, Clustering–type and queuing–type models, Distributional limit theorems, Refined approximations, Modular analysis, Renewal theory, Risk theory, ergodic Markov and semi–Markov systems.

1. Introduction

In his review [46] of Petrov’s book [80], Kesten111Harry Kesten (1931–2019), the Goldwin Smith Professor Emeritus of Mathematics, “whose insights advanced the modern understanding of probability theory and its applications” (quoted from obituary by Matt Hayes). wrote that “in the period 1920–1940 research in probability theory was virtually synonymous with the study of sums of independent random variables”, but “already in the late thirties attention started shifting and probabilists became more interested in Markov chains, continuous time processes, and other situations in which the independence between summands no longer applies, and today the center of gravity of probability theory has moved away from sums of independent random variables.”

Nevertheless, Kesten continues, “much work on sums of independent random variables continues to be done, partly for their aesthetic appeal and partly for the technical reason that many limit theorems, even for dependent summands, can be reduced to the case of independent summands by means of various tricks. The best known of these tricks is the use of regeneration points, or ‘Dœblin’s trick’. In large part the fascination of the subject is due to the fact that applications and the ingenuity of mathematicians continue to give rise to new questions.” Regarding applications, Kesten mentioned renewal theory, the theory of optimal stopping, invariance principles and functional limit theorems, fluctuation theory, and so on.

To summarize, Kesten argues that “in all the years that these newer phenomena were being discovered, people also continued working on more direct generalizations and refinements of the classical limit theorems \dots Many of the proofs require tremendous skill in classical analysis \dots and many of the results, such as the Berry–Esseen estimate, the Edgeworth expansion and asymptotic results for large deviations, are important for statistics and theoretical purposes.”

The paper is an overview that fits into the scheme outlined by Kesten. It is focussed on refined “distributional” (or “intrinsically analytic”222The terms “intrinsically analytic” and “measure–theoretic” applied to limit theorems are introduced in [19]. The former is equivalent (see, e.g., [12]) to “distributional”, or “weak”., or “weak”) limit theorems for compound sums which are a natural extension of ordinary sums. They are known as the key object of renewal theory and the core component of Dœblin’s dissection of an ordinary sum defined on a Markov chain. By “refined” limit theorems we mean333The results for large deviations, although interesting and available, are not presented here due to the space limitation. Berry–Esseen’s estimates and Edgeworth’s expansions mentioned by Kesten, in contrast to approximations without assessing their accuracy.

In contrast to “measure–theoretic” limit theorems, for which Kolmogorov’s axiomatization is indispensable and which border on (or lie within) the theory of random processes, “distributional” limit theorems, including those for dependent random variables, are known long before the axiomatization. For example, seeking in the 1900s, long before Kolmogorov’s advance, to demonstrate that independence is not a necessary condition in “distributional” weak law of large numbers (WLLN) and central limit theorem (CLT), Markov introduced the Markov chains.

Without questioning the value of “measure–theoretic” limit theorems, such as the strong law of large numbers (SLLN) and the law of iterated logarithm (LIL), or, going further, the value of the theory of random processes, we note that “distributional” rather than “measure–theoretic” limit theorems are really needed for statistics, cluster analysis, etc., where an infinite number of observations, clusters, renewals, people in queue, and so on, never occur. Even in Bernoulli’s classical scheme, statistical inference is based (see detailed discussion in [18]) on WLLN and CLT with its refinements, and not on SLLN and LIL.

Dealing with complex compound sums444See classification of compound sums below. Modular analysis is not needed for simple compound sums., we focus on modular analysis. Speaking of “Dœblin’s trick” which reduces “many limit theorems, even for dependent summands” to the case of independent ones, Kesten meant this approach. Bearing in mind the ladder technique, say “Blackwell’s trick” (see [13]), the modular analysis is even more widely used. We will expand it to compound sums with general modular structure, but note that (see, e.g., [39], [42], [65], [83], [87]) this is merely one, rather than the only and the best, method for studying such sums.

Technically, “Dœblin’s trick” (or its counterpart “Blackwell’s trick”) is not only the use of regeneration (or ladder) points. It implies (see, e.g., [20]) the use of Kolmogorov’s inequality for maximum of partial sums in order to move from Dœblin’s (or Blackwell’s) dissection to ordinary sum of independent modular summands. We call this approach “basic technique”.

Delving deeper, the “basic technique” in its original form fails in local theorems and yields no “refined” approximations, such as Berry–Esseen’s estimates and Edgeworth’s expansions. To be specific, dealing with a local limit theorem for classical Markov chains, Kolmogorov (see [49]) pointed that “Dœblin’s method in its original form” is suitable “only for proving integral theorems”. We quote555In our translation from Russian. Apparently, this article has not been translated in due course from Russian into English. more from [49], § 2: “\dots the local theorems that form the main content of the present paper will be obtained \dots with the help of some strengthening of the method that was developed by Dœblin for proving the integral limit theorem in the case of an infinite number of states. In order to make the development of the method clear, we single out \dots a brief exposition of Dœblin’s method in its original form, in which it is suitable only for proving integral theorems.”

In brief, Kolmogorov’s advance in [49] was due to a straightforward use of the total probability formula, or, according to his terminology, “basic identity”.

Further progress was made thirty years later in [14] and in a series of papers (see, e.g., [15], [43], [55][60]) that followed: Berry–Esseen’s estimates and Edgeworth’s expansions in CLT for recurrent Markov chain were obtained by using the total probability rule and the auxiliary multidimensional limit theorems for sums of independent random vectors. To distinguish this approach from “basic technique”, or “Dœblin’s method in its original form”, we call it “advanced (modular, if switching to modular summands is needed) technique”.

Being one of the main pillars of the advanced technique, the limit theorems for sums of independent random vectors (see, e.g., [11], [31][33]) are close to perfection. In the advanced technique, analytical complexity, which never disappears but flows from one form to another, is split into the use of, firstly, the auxiliary theorems where (we quote from Preface of [11]) “precision and generality go hand in hand” and, secondly, tedious but fairly standard classical analysis, e.g., approximation of integral sums by appropriate integrals and evaluation of remainder terms.

Further presentation is arranged as follows. Section 2 is devoted to genesis and classification of compound sums. In Section 3, we address elementary and refined renewal theorems and refined limit theorems for cumulated rewards. Having formulated them as limit theorems for simple compound sums, we outline their proof using advanced technique instead of renewal equation, Laplace transforms, and Tauberian theorems commonly used in renewal theory.

In Section 4, we address complex compound sums “with irregular summation and independence” related to the ruin problem of collective risk theory. To get (refined) normal approximation when such sums are proper, we use the modular analysis based on Blackwell’s ladder idea. To get (refined) quasi–normal approximation when such sums are defective, we resort to associated random variables. In this context, we touch upon the inverse Gaussian approximation which deviates somewhat from the main topic of this paper, but demonstrates that there are more sophisticated methods than those related to asymptotic normality.

In Section 5, seeking to use the full force of modular analysis combined with advanced technique, we focus on compound sums with general modular structure. Particular cases of this framework are complex compound sums “with irregular summation and dependence”, in particular Markov dependence. The only case, although of significant practical interest, where great impediments for asymptotic analysis arise is defective compound sums “with irregular summation and dependence”. This case, which is tackled by means of computer–intensive analysis, is briefly discussed in Section 6. A concluding remark is made in Section 7.

The article, being a review, does not contain complete proofs. Most results are formulated in a non–strict manner, although the necessary references are given everywhere. Striving for clarity, great emphasis is placed on discussing the main ideas of the proofs. Overall, the aims of the advance described in this review are the same as Loève stated in [53] writing on limit theorems of probability theory: (i) to simplify proofs and forge general tools out of the special ones, (ii) to sharpen and strengthen results, (iii) to find general notions behind the results obtained and to extend their domains of validity.

2. Genesis and classification of compound sums

It was observed (see [69]) that “on the Continent, if people are waiting at a bus–stop, they loiter around in a seemingly vague fashion. When the bus arrives, they make a dash for it \dots An Englishman, even if he is alone, forms an orderly queue of one.” In terms of mathematical modeling, this is a “clustering–type” model opposed to a “queuing–type” model. In the former, there is essentially no flow of time, and the objects or events of interest are scattered in space. In the latter, there is a flow of time, and the objects or events of interest form an ordered queue.

In models of both types, the core is the compound sum

SNt=i=1NtXi{S}_{N_{t}}=\sum_{i=1}^{N_{t}}X_{i}

with basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, where t>0t>0 and

Nt=inf{n1:i=1nTi>t}N_{t}=\inf\Big{\{}n\geqslant 1:\sum_{i=1}^{n}T_{i}>t\Big{\}} (2.1)

or ++\infty, if the set is empty666Recall that inf=+\inf\emptyset=+\infty, sup=0\sup\emptyset=0.. When Nt=+N_{t}=+\infty, the sum SNt{S}_{N_{t}} is set equal to ++\infty. For brevity, the real–valued random variables XiX_{i} are called primary components (pp–components) and TiT_{i} secondary components (ss–components) of the basis.

If the basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, consists of independent random vectors, then SNt{S}_{N_{t}} is “with independence”. Otherwise it is “with dependence”. Examples of the latter are often found (see, e.g., [78]) in various models with Markov modulation. If ss–components are positive, then SNt{S}_{N_{t}} is “with regular summation”. Otherwise, it is “with irregular summation”.

Apparently, compound sums “with regular summation and independence” fully and distinctly came forth in “queuing–type” renewal theory and its ramifications. Compound sums “with irregular summation and independence” appeared (see, e.g., [66]) in ruin problem and (see, e.g., [41]) in many settings which involve random walks. On the other hand, Dœblin’s dissection, i.e., switching from ordinary sum defined on a Markov chain to modular simple compound sum “with regular summation and independence”, whose basis consists of intervals between regeneration points and increments at these intervals, is another source of compound sums.

The complexity of “irregular summation” is rather transparent: the sums Vn=i=1nTi{V}_{n}=\sum_{i=1}^{n}T_{i}, n=1,2,n=1,2,\dots, are not progressively increasing, whence {Nt=n}\{N_{t}=n\} is equal to {V1t,,Vn2t,Vn1t<Vn}\{{V}_{1}\leqslant t,\dots,{V}_{n-2}\leqslant t,{V}_{n-1}\leqslant t<{V}_{n}\}, rather than {Vn1t<Vn}\{{V}_{n-1}\leqslant t<{V}_{n}\}, as would be for “regular summation”.

In “dependence complexity” case, the model has to be fleshed out. In the search of non–trivial examples which illustrate the general results, we focus on Markov dependence and its extensions, but in fact any type of dependence which allows an embedded modular structure with independent modules suits us well. In terms of random processes, much of this topic is based on successive “starting over”, as time goes on, commonly called “regeneration” or (see [47], [78]) “regenerative phenomena”. With this approach, many cases that go beyond the scope of Markov theory fall within the scope of this paper.

Remark 2.1.

Beyond the scope of this paper are the sums of the following two types. First, the random sums SN{S}_{N} with NN and X1,X2,X_{1},X_{2},\dots not necessarily independent of each other, but with NN not of the form (2.1) or the similar. Second, the random sums SN=i=1NXi{S}_{N}=\sum_{i=1}^{N}X_{i} with independence between NN and X1,X2,X_{1},X_{2},\dots. The former were introduced in [6], [81], [82], and investigated, e.g., in [41]. The latter were introduced in [84] and thoroughly studied (see, e.g., [38]) in many works.∎

2.1. Proper and defective compound sums

{bundle}

Compound sum \chunk /\mathcal{R\hskip-0.5pt/I}  sum \chunk{bundle} /\mathcal{I\hskip-0.5pt/I}  sum \chunk proper /\mathcal{I\hskip-0.5pt/I}  sum \chunk defective /\mathcal{I\hskip-0.5pt/I}  sum \chunk /𝒟\mathcal{R\hskip-0.5pt/D}  sum \chunk{bundle} /𝒟\mathcal{I\hskip-0.5pt/D}  sum \chunk proper /𝒟\mathcal{I\hskip-0.5pt/D}  sum \chunk defective /𝒟\mathcal{I\hskip-0.5pt/D}  sum

Figure 1. Classification of compound sums.

The classification of compound sums, where shorthand for “compound sum with regular summation and independence” is “/\mathcal{R\hskip-0.5pt/I}  sum”, for “compound sum with irregular summation and independence” is “/\mathcal{I\hskip-0.5pt/I}  sum”, for “compound sum with regular summation and dependence” is “/𝒟\mathcal{R\hskip-0.5pt/D}  sum”, and for “compound sum with irregular summation and dependence” is “/𝒟\mathcal{I\hskip-0.5pt/D}  sum”, is shown in Fig. 1. Under natural regularity conditions, compound sums “with regular summation” are always proper, while compound sums “with irregular summation” are either proper, or defective777Recall (see [37]) that a random variable is defective, with defect 1p1-p, if it takes the value \infty with probability p>0p>0. Otherwise, it is proper..

The following example shows that finding the defect of /\mathcal{I\hskip-0.5pt/I}  sum is not easy.

Example 2.1.

Let (Xi,Yi)(X_{i},Y_{i}), i=1,2,i=1,2,\dots, be a subsidiary sequence of i.i.d.  random vectors, whose components are positive and independent of each other. For c>0c>0 and Ti=YicXiT_{i}=Y_{i}-cX_{i}, SNt=i=1NtXi{S}_{N_{t}}=\sum_{i=1}^{N_{t}}X_{i} is positive /\mathcal{I\hskip-0.5pt/I}  sum with basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots. While the components of the subsidiary sequence are independent of each other, the components of the basis are dependent on each other. If XiX_{i} and YiY_{i}, i=1,2,i=1,2,\dots, are exponentially distributed with positive parameters ϱ\varrho and ρ\rho, then (see, e.g., [65], Theorem 5.8)

𝖯{SNtx}=𝖯{SNt<}1π0πf(z,t,x)𝑑z,\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}}=\mathsf{P}\big{\{}{S}_{N_{t}}<\infty\big{\}}-\frac{1}{\pi}\int_{0}^{\pi}f(z,t,x)\,dz, (2.2)

where

𝖯{SNt<}={1,ϱ/(cρ)1,ϱcρexp{t(cρϱ)/c},ϱ/(cρ)<1\mathsf{P}\big{\{}{S}_{N_{t}}<\infty\big{\}}=\begin{cases}1,&\varrho/(c\rho)\geqslant 1,\\[8.0pt] \dfrac{\varrho}{c\rho}\,\exp\big{\{}-t(c\rho-\varrho)/c\big{\}},&\varrho/(c\rho)<1\end{cases} (2.3)

and

f(z,t,x)=\displaystyle f(z,t,x)= (ϱ/(cρ))(1+ϱ/(cρ)2ϱ/(cρ)cosz)1\displaystyle\,(\varrho/(c\rho))(1+\varrho/(c\rho)-2\sqrt{\varrho/(c\rho)}\cos z)^{-1} (2.4)
×exp{tρ(ϱ/(cρ)cosz1)\displaystyle\times\exp\big{\{}t\rho\,(\sqrt{\varrho/(c\rho)}\cos z-1)
xϱ(cρ/ϱ)(1+ϱ/(cρ)2ϱ/(cρ)cosz)}\displaystyle-x\varrho(c\rho/\varrho)(1+\varrho/(c\rho)-2\sqrt{\varrho/(c\rho)}\cos z)\big{\}}
×(cos(tρϱ/(cρ)sinz)cos(tρϱ/(cρ)sinz+2z)).\displaystyle\times(\cos(t\rho\sqrt{\varrho/(c\rho)}\sin z)-\cos(t\rho\sqrt{\varrho/(c\rho)}\sin z+2z)).

Consequently, SNt{S}_{N_{t}} is proper if ϱ/(cρ)1\varrho/(c\rho)\geqslant 1, which is equivalent to cc=𝖤Y/𝖤Xc\leqslant c^{\ast}=\mathsf{E}{Y}/\mathsf{E}{X}, and defective otherwise. Defect of SNt{S}_{N_{t}} is yielded in (2.3). ∎

Although the expression (2.3) for defect of SNt{S}_{N_{t}} is complicated888This is a particular case of Cramér’s famous result on the probability of ultimate ruin., the conditions under which SNt{S}_{N_{t}} is proper or defective, are simple. On the one hand, if 𝖤T1>0\mathsf{E}T_{1}>0, then (see Theorem 2 in [37], Chapter XII) the ordinary sums Vn{V}_{n} increase a.s.  to ++\infty, as nn\to\infty. Consequently, for any tt the summation limit NtN_{t} is a.s.  finite and SNt{S}_{N_{t}} is a.s.  less than ++\infty, whence proper. On the other hand, if 𝖤T1<0\mathsf{E}T_{1}<0, then the ordinary sums Vn{V}_{n} decrease a.s.  to -\infty, as nn\to\infty. With a positive probability, the maximum M=maxn𝖭VnM=\max_{n\in\mathsf{N}}{V}_{n} is finite and less than t>0t>0 large enough, whence NtN_{t} and SNt{S}_{N_{t}} are defective.

2.2. Renewal theory and species of compound sums

In renewal theory, pp–components XiX_{i}, i=1,2,i=1,2,\dots, (typically positive) are called “rewards in the moments of renewals” and ss–components TiT_{i}, i=1,2,i=1,2,\dots, (always positive) are called “intervals between successive renewals”. This interpretation yields a variety of species of /\mathcal{R\hskip-0.5pt/I}  sums. In particular, the total number of renewals before time tt, defined as N~t=sup{n1:Vnt}\widetilde{N}_{t}=\sup\{n\geqslant 1:{V}_{n}\leqslant t\}, or 0, if T1>tT_{1}>t, comes to the fore. Focus shifts from NtN_{t} to N~t\widetilde{N}_{t}, which coincides with N^t=Nt1\widehat{N}_{t}=N_{t}-1 trivially related to NtN_{t}.

When moving from relatively simple /\mathcal{R\hskip-0.5pt/I}  sums to more complicated /\mathcal{I\hskip-0.5pt/I} and /𝒟\mathcal{I\hskip-0.5pt/D}  sums, both of them with “irregular summation”, the difference between altered (with summation limit N~t\widetilde{N}_{t}) and non–altered (with summation limit NtN_{t}) compound sums becomes non–trivial: N~t\widetilde{N}_{t} is no more trivially related to NtN_{t} and the altered compound sums SN~t=i=1N~tXi{S}_{\widetilde{N}_{t}}=\sum_{i=1}^{\widetilde{N}_{t}}X_{i} and SN^t=i=1N^tXi{S}_{\widehat{N}_{t}}=\sum_{i=1}^{\widehat{N}_{t}}X_{i} are not bounded to coincide.

As for /𝒟\mathcal{R\hskip-0.5pt/D}  sums with Markov dependence, which generalize /\mathcal{R\hskip-0.5pt/I}  sums in the same way as Markov renewal theory generalizes renewal theory, dependence is introduced by embedded Markov chain {ξi}i0\{\xi_{i}\}_{i\geqslant 0}. In such /𝒟\mathcal{R\hskip-0.5pt/D}  sums, ss–components TiT_{i}, i=1,2,i=1,2,\dots, are the time intervals between jumps of {ξi}i0\{\xi_{i}\}_{i\geqslant 0} and pp–components XiX_{i}, i=1,2,i=1,2,\dots, which depend on {ξi}i0\{\xi_{i}\}_{i\geqslant 0}, are rewards in the moments of renewals.

In this way, the variety of species of compound sums echoes the diversity of renewal processes. There are999We refer to [21], Section 2.2, or [22], Section 9.2. There are terminological collisions, e.g., in [45], Chapter 5, Section 7, this process is called delayed renewal process. modified, equilibrium, and ordinary renewal processes. If T1T_{1} is distributed differently than Ti=𝑑TT_{i}\overset{d}{=}T, i=2,3,i=2,3,\dots, then the renewal process is called modified. When all intervals between renewals, including the first one, are identically distributed, the renewal process is called ordinary. Furthermore, if

𝖯{T1x}=1𝖤T0x(1FT(z))𝑑z,x>0,\mathsf{P}\big{\{}T_{1}\leqslant x\big{\}}=\frac{1}{\mathsf{E}T}\int_{0}^{x}\big{(}1-F_{T}(z)\big{)}\,dz,\quad x>0,

where FTF_{T} denotes c.d.f.  of TT, then it is called equilibrium, or stationary.

Regarding compound sums, modified are those in which a finite number of elements of the basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, e.g., the first element (T1,X1)(T_{1},X_{1}), differ from all the others. Unlike in renewal theory which models technical repairable systems where the first repair does not necessarily occur at the starting time zero, this modification looks awkward for /\mathcal{R\hskip-0.5pt/I}  sums: it boils down to a very special case of non–identically distributed summands. The similar modification of /𝒟\mathcal{R\hskip-0.5pt/D}  sums with Markov dependence comes down to the choice of the initial distribution of the embedded Markov chain {ξi}i0\{\xi_{i}\}_{i\geqslant 0} and is much more sensible.

It is noteworthy that, under natural regularity conditions, switching within the species of compound sums, e.g., from altered to unaltered compound sums and vice versa, does not affect the main–term approximations, but affects refinements such as Edgeworth’s expansions.

3. Renewal theory and limit theorems for 𝓡/𝓘\boldsymbol{\mathcal{R\hskip-0.5pt/I}}  sums

Founded (see [34], [35]) in the 1940s as a theoretical insight into technical repairable systems, the classical renewal theory is focussed on “queuing–type” models. The term “renewal”, which displaced the formerly used term “industrial replacement” coined (see [54]) in the 1930s by Lotka, is widely used nowadays, but was not the main or default term even in Feller’s seminal paper [36] dated 1949.

In 1948, Doob [30] noted that “renewal theory is ordinarily reduced to the theory of certain types of integral equations \dots However, it is to be expected that a treatment in terms of the theory of probability, which uses the modern developments of this theory, will shed new light on the subject.” Feller agreed (see [37], Chapter VI) that in renewal theory “analytically, we are concerned merely with sums of independent positive variables.”

3.1. First appearance of the advanced technique

In words, the renewal function101010The asterisk denotes the convolution operator. 𝖴(t)=n=0FTn(t)\mathsf{U}(t)=\sum_{n=0}^{\infty}F_{T}^{*n}(t), t>0t>0, where FTF_{T} denotes c.d.f.  of a positive random variable TT, is the expected number of renewals in the time interval (0,t)(0,t), with the origin counted as renewal epoch111111In [22], [45], the renewal function is defined as 𝖧(t)=𝖴(t)1\mathsf{H}(t)=\mathsf{U}(t)-1, t>0t>0 (or 𝖤N^t\mathsf{E}\widehat{N}_{t}), which differs from 𝖴(t)\mathsf{U}(t) by one. In [37], the function 𝖴(t)\mathsf{U}(t), t>0t>0, is called the (ordinary) renewal process. In [22], [45], the renewal process is referred to as the continuous–time random process {N^t}t0\{\widehat{N}_{t}\}_{t\geqslant 0}, i.e., the random number of renewals in the time interval (0,t)(0,t), with the origin not being counted as a renewal epoch.. As a formula, this is 𝖴(t)=1+𝖤N^t\mathsf{U}(t)=1+\mathsf{E}\widehat{N}_{t}. Together with altered /\mathcal{R\hskip-0.5pt/I}  sum SN^t{S}_{\widehat{N}_{t}}, called in renewal theory “cumulated rewards”, it is built on /\mathcal{R\hskip-0.5pt/I}  basis (Ti,Xi)=𝑑(T,X)(T_{i},X_{i})\overset{d}{=}(T,X), i=1,2,i=1,2,\dots, with i.i.d.  components and TT positive.

The elementary renewal theorem (see, e.g., [41]), states that for 0<𝖤T0<\mathsf{E}\,T\leqslant\infty the approximation121212When 𝖤T=\mathsf{E}\,T=\infty, the ratio t/𝖤Tt/\mathsf{E}\,T is replaced by 0.

𝖤N^t=t𝖤T+o(t),t,\mathsf{E}\widehat{N}_{t}=\frac{t}{\mathsf{E}\,T}+o(t),\quad t\to\infty, (3.1)

holds. The corresponding expansion up to vanishing term is

𝖤N^t=t𝖤T+𝖣T(𝖤T)22(𝖤T)2+o(1),t,\mathsf{E}\widehat{N}_{t}=\frac{t}{\mathsf{E}\,T}+\frac{\mathsf{D}T-(\mathsf{E}\,T\,)^{2}}{2\,(\mathsf{E}\,T\,)^{2}}+o(1),\quad t\to\infty, (3.2)

called refined elementary renewal theorem.

The elementary renewal theorem is widely known: see, e.g., Equality (3) in [21], Section 4.2, or Equality (17) in [22], Section 9.2, or main formula in point (b) in [45], Chapter 5, Section 6. In [22], p. 345, it is noted that “a rigorous proof \dots is possible under very weak assumptions about the distribution of TT, but requires difficult Tauberian arguments and will not be attempted here”. It can also be obtained (with some caution with regard to the definition of renewal function) from Theorem 1 in [37], Chapter XI, the proof of which is based on the renewal equation.

The refined elementary renewal theorem and the similar results for higher–order power moments of N^t\widehat{N}_{t} and SN^t{S}_{\widehat{N}_{t}}, including expansions up to vanishing terms, were studied (see, e.g., [21], Section 4.5, and [3]) by various methods.

We outline a proof (see details in [65], Chapter 4) that does not compete with the proofs mentioned above in the sense of elegance or minimality of conditions. The advantage of this proof, based on the use of refined CLT and consisting of steps  A–F, is that it will be routinely carried over to much more complex settings.

Step  A:  use of fundamental identity.  The identity

𝖤N^t=\displaystyle\mathsf{E}\widehat{N}_{t}= n=1n𝖯{N^t=n}=n=1n𝖯{i=1nTit<i=1n+1Ti}\displaystyle\,\sum_{n=1}^{\infty}n\;\mathsf{P}\big{\{}\widehat{N}_{t}=n\big{\}}=\sum_{n=1}^{\infty}n\;\mathsf{P}\bigg{\{}\sum_{i=1}^{n}T_{i}\leqslant t<\sum_{i=1}^{n+1}T_{i}\bigg{\}} (3.3)
=\displaystyle= n=1n(FTn(t)FT(n+1)(t))\displaystyle\,\sum_{n=1}^{\infty}n\big{(}F_{T}^{*n}(t)-F_{T}^{*(n+1)}(t)\big{)}

is obvious. If the probability density function (p.d.f.) fTf_{T} exists (we assume this for simplicity of presentation), then (3.3) can be rewritten as

𝖤N^t=n=1n0tfTn(tz)𝖯{Tn+1>z}𝑑z.\displaystyle\mathsf{E}\widehat{N}_{t}=\sum_{n=1}^{\infty}n\int_{0}^{t}f_{T}^{*n}(t-z)\,\mathsf{P}\big{\{}T_{n+1}>z\big{\}}\,dz. (3.4)

Step  B:  reduction of range of summation.  At this step, we cut out from the sum on the right side of (3.3) those terms that correspond to small nn (i.e., n<ntn<n_{t} with properly selected ntn_{t}\to\infty, as tt\to\infty). This is intuitively clear: for large tt, the random variable N^t\widehat{N}_{t} is likely to be large. Therefore, the probabilities 𝖯{N^t=n}\mathsf{P}\big{\{}\widehat{N}_{t}=n\big{\}} for small nn are small. Technically, it is done using the well–known probability inequalities, such as Markov’s inequality.

Step  C:  reduction of range of integration.  At this step, relying on the moment conditions, we cut out from the integral in (3.4) that part that corresponds to large zz (i.e., z>ztz>z_{t} with properly selected ztz_{t}\to\infty, as tt\to\infty). This is also intuitively clear: any single random variable, including Tn+1T_{n+1}, is small compared to the sum i=1nTi\sum_{i=1}^{n}T_{i}, as nn is sufficiently large.

Step  D:  application of refined CLT.  Writing μT=𝖤T\mu_{T}=\mathsf{E}\,T, σT2=𝖣T\sigma_{T}^{2}=\mathsf{D}\,T and switching to standardized random variables T~i=(TiμT)/σT\tilde{T}_{i}=(T_{i}-\mu_{T})/\sigma_{T}, i=1,2,i=1,2,\dots, we have

fTn(tz)=1σTnfn1/2i=1nT~i(λT,n(tz)),f_{T}^{*n}(t-z)=\frac{1}{\sigma_{T}\sqrt{n}}\,f_{\,n^{-1/2}\sum_{i=1}^{n}\tilde{T}_{i}}\big{(}\lambda_{T,n}{(t-z)}\big{)},

where λT,n(tz)=(tzμTn)/(σTn)\lambda_{T,n}{(t-z)}=(t-z-\mu_{T}n)/(\sigma_{T}\sqrt{n}). Assuming that natural moment conditions are satisfied, we apply Berry–Esseen’s estimate (see Theorem 11 in [80], Chapter VII, § 2) in the proof of (3.1) and Edgeworth’s expansion (see Theorem 17 in [80], Chapter VII, § 3) in the proof of (3.2). In both cases, these results are taken with non–uniform remainder terms.

Step  E:  analysis of approximating term.  In the proof of (3.1), where Berry–Esseen’s estimate was used, the approximating term is

t=1σTn>ntn1/20ztφ(0,1)(λT,n(tz))𝖯{T>z}𝑑z,\mathcal{I}_{t}=\frac{1}{\sigma_{T}}\sum_{n>n_{t}}n^{1/2}\int_{0}^{z_{t}}\varphi_{\left({0},{1}\right)}\big{(}\lambda_{T,n}{(t-z)}\big{)}\,\mathsf{P}\big{\{}T>z\big{\}}\,dz,

where φ(0,1)\varphi_{\left({0},{1}\right)} denotes p.d.f.  of standard normal distribution. In the proof of (3.2), where Edgeworth’s expansion was used, the approximating term is the sum of

t=\displaystyle\mathcal{I}_{t}^{\,{\dagger}}= 1σTn>ntn0ztφ(0,1)(λT,n(tz))𝖯{T>z}𝑑z\displaystyle\,\frac{1}{\sigma_{T}}\sum_{n>n_{t}}\sqrt{n}\int_{0}^{z_{t}}\varphi_{\left({0},{1}\right)}(\lambda_{T,n}{(t-z)})\,\mathsf{P}\big{\{}T>z\big{\}}\,dz
t=\displaystyle\mathcal{I}_{t}^{\,{\ddagger}}= 1σT𝖤(T3)6σT3n>nt0zt(λT,n3(tz)3λT,n(tz))\displaystyle\,\frac{1}{\sigma_{T}}\frac{\mathsf{E}(T^{3})}{6\,\sigma_{T}^{3}}\sum_{n>n_{t}}\int_{0}^{z_{t}}\,\big{(}\lambda_{T,n}^{3}{(t-z)}-3\,\lambda_{T,n}{(t-z)}\big{)}
×φ(0,1)(λT,n(tz))𝖯{T>z}dz.\displaystyle\times\varphi_{\left({0},{1}\right)}\big{(}\lambda_{T,n}{(t-z)}\big{)}\,\mathsf{P}\big{\{}T>z\big{\}}\,dz.

Using Riemann summation formula with nodal points λT,n(t)<λT,n1(t)<<λT,2(t)<λT,1(t)\lambda_{T,n}{(t)}{}<\lambda_{T,n-1}{(t)}{}<\dots<\lambda_{T,2}{(t)}{}<\lambda_{T,1}{(t)}{}, we seek to reduce t\mathcal{I}_{t} to t/𝖤T{t}/{\mathsf{E}\,T} with the required accuracy, and t+t\mathcal{I}_{t}^{\,{\dagger}}+\mathcal{I}_{t}^{\,{\ddagger}} to t/𝖤T+(𝖣T(𝖤T)2)/(2(𝖤T)2)t/\mathsf{E}\,T+(\mathsf{D}T-(\mathsf{E}\,T)^{2})/(2\,(\mathsf{E}\,T)^{2}) with the required accuracy. This is a tedious but fairly standard classical analysis left to the reader.

Step  F:  analysis of remainder term.  In the proof of (3.1), the remainder term is

t=C1n>ntδn0zt(1+|λT,n(tz)|)3𝖯{T>z}𝑑z.\mathcal{R}_{t}={C}_{1}\sum_{n>n_{t}}\delta_{n}\int_{0}^{z_{t}}\big{(}1+|\,\lambda_{T,n}{(t-z)}\,|\,\big{)}^{-3}\,\mathsf{P}\big{\{}T>z\big{\}}\,dz.

In the proof of (3.2), the remainder term is

t=\displaystyle\mathcal{R}_{t}= C2n>ntδnn0zt(1+|λT,n(tz)|)4𝖯{T>z}𝑑z,\displaystyle\,{C}_{2}\sum_{n>n_{t}}\frac{\delta_{n}}{\sqrt{n}}\int_{0}^{z_{t}}\big{(}1+\big{|}\,\lambda_{T,n}{(t-z)}\,\big{|}\,\big{)}^{-4}\,\mathsf{P}\big{\{}T>z\big{\}}\,dz,

with δn0\delta_{n}\to 0, as nn\to\infty. The former is o(t)o(t), tt\to\infty. The latter is o(1)o(1), tt\to\infty, which is shown by direct analytical methods. This is also a tedious but fairly standard classical analysis left to the reader.

3.2. Refined limit theorems for cumulated rewards

Moving from renewal function (or power moments of N^t\widehat{N}_{t}) to the distribution of altered /\mathcal{R\hskip-0.5pt/I}  sum SN^t=i=1N^tXi{S}_{\widehat{N}_{t}}=\sum_{i=1}^{\widehat{N}_{t}}X_{i}, called in renewal theory “cumulated rewards”, we use the shorthand notation μX=𝖤X\mu_{X}=\mathsf{E}X, σX2=𝖣X\sigma_{X}^{2}=\mathsf{D}X, ϰXT(i,j)=𝖤((XμX)i(TμT)j)\varkappa_{XT}^{\,(i,j)}=\mathsf{E}((X-\mu_{X})^{i}(T-\mu_{T})^{j}) for ii and jj integers, and ϰXT=ϰXT(1,1)\varkappa_{XT}=\varkappa_{XT}^{\,(1,1)}. We put

MS=μXμT1,DS2=𝖤(μXTμTX)2μT3.M_{{S}}{\,\!}=\mu_{X}\mu_{T}^{-1},\quad D_{{S}}^{2}{\,}=\mathsf{E}\,(\mu_{X}T-\mu_{T}X)^{2}\mu_{T}^{-3}.

By Φ(0,1)(x)\varPhi_{\left({0},{1}\right)}(x), we denote c.d.f.  of standard normal distribution.

Theorem 3.1 (Berry–Esseen’s estimate).

If p.d.f.  of TT is bounded above by a finite constant131313In Theorem 3.1, this condition is excessive. We repeat that we do not strive for maximum generality (or rigor) in this presentation., DS2>0D_{{S}}^{2}{\,}>0, and 𝖤(T3)<\mathsf{E}(T^{3})<\infty, 𝖤(X3)<\mathsf{E}(X^{3})<\infty, then

supx𝖱|𝖯{SN^tMStxDSt1/2}Φ(0,1)(x)|=O(t1/2),t.\sup_{x\in\mathsf{R}}\,\Big{|}\,\mathsf{P}\big{\{}{S}_{\widehat{N}_{t}}-M_{{S}}{\,t}\leqslant x\,D_{{S}}{\,t^{1/2}}\big{\}}-\varPhi_{\left({0},{1}\right)}(x)\,\Big{|}=O\big{(}t^{-1/2}\big{)},\quad t\to\infty.
Theorem 3.2 (Edgeworth’s expansion).

If the conditions of Theorem 3.1 are satisfied and 𝖤(Xk)<\mathsf{E}(X^{k})<\infty, 𝖤(Tk)<\mathsf{E}(T^{k})<\infty, then there exist polynomials 𝖰r(x)\mathsf{Q}_{\,r}(x), r=1,2,,k3r=1,2,\dots,k-3, of degree 3r13r-1, such that

supx𝖱|𝖯{SN^tMStxDSt1/2}Φ(0,1)(x)\displaystyle\sup_{x\in\mathsf{R}}\bigg{|}\,\mathsf{P}\big{\{}{S}_{\widehat{N}_{t}}-M_{{S}}{\,t}\leqslant xD_{{S}}{\,t^{1/2}}\big{\}}-\varPhi_{\left({0},{1}\right)}(x)
r=1k3𝖰r(x)tr/2φ(0,1)(x)|=O(t(k2)/2),t.\displaystyle\hskip 90.0pt-\sum_{r=1}^{k-3}\frac{\mathsf{Q}_{\,r}(x)}{t^{r/2}}\,\,\varphi_{\left({0},{1}\right)}(x)\bigg{|}=O\big{(}t^{-(k-2)/2}\big{)},\quad t\to\infty.

In particular,

𝖰 1(x)=16(K(3,0)𝖧2(x)+3I1),\mathsf{Q}_{\,1}(x)=-\frac{1}{6}\,\big{(}K_{(3,0)}\mathsf{H}_{2}(x)+3\hskip 0.2pt{I}_{1}\big{)},

where 𝖧2(x)=x21\mathsf{H}_{2}(x)=x^{2}-1 is Chebyshev–Hermite’s polynomial,

K(3,0)=\displaystyle K_{(3,0)}= ((ϰXT(3,0)3σX2ϰXTμT1+6σX2σT2μTμX26σT2ϰXTμX2μT3)μT1\displaystyle\,\big{(}\big{(}\varkappa_{XT}^{\,(3,0)}-3\sigma_{X}^{2}\varkappa_{XT}\mu_{T}^{-1}+6\sigma_{X}^{2}\sigma_{T}^{2}\mu_{T}\mu_{X}^{-2}-6\sigma_{T}^{2}\varkappa_{XT}\mu_{X}^{2}\mu_{T}^{-3}\big{)}\,\mu_{T}^{-1}
3(ϰXT(2,1)2ϰXT2μT1+σX2σT2μT1)μTμX2\displaystyle-3\,\big{(}\varkappa_{XT}^{\,(2,1)}-2\varkappa_{XT}^{2}\mu_{T}^{-1}+\sigma_{X}^{2}\sigma_{T}^{2}\mu_{T}^{-1}\big{)}\,\mu_{T}\mu_{X}^{-2}
+3(ϰXT(1,2)σT2ϰXTμT1)μX2μT3\displaystyle+3\,\big{(}\varkappa_{XT}^{\,(1,2)}-\sigma_{T}^{2}\varkappa_{XT}\mu_{T}^{-1}\big{)}\,\mu_{X}^{2}\mu_{T}^{-3}
(ϰXT(0,3)3σT4μT1)μX3μT4)DS3,\displaystyle-\big{(}\varkappa_{XT}^{\,(0,3)}-3\sigma_{T}^{4}\mu_{T}^{-1}\big{)}\,\mu_{X}^{3}\mu_{T}^{-4}\big{)}\,D_{{S}}^{-3}{\,},

and I1=μX(σT2μT2+1)DS1{I}_{1}=\mu_{X}\big{(}\sigma_{T}^{2}\,\mu_{T}^{-2}+1\big{)}\,D_{{S}}^{-1}{\,}.

Both Theorems 3.1 and 3.2 are proved (see details in [65], Chapter 4) by means of practically the same advanced technique which was used in the proof of (3.1) and (3.2), respectively. It consists of steps A–F, which we will draw.

Step  A:  use of fundamental identity.  Equalities (3.3) and (3.4) are replaced by

𝖯{SN^tx}=\displaystyle\mathsf{P}\big{\{}{S}_{\widehat{N}_{t}}\leqslant x\big{\}}= n=10t𝖯{i=1nXix|i=1nTi=tz}\displaystyle\,\sum_{n=1}^{\infty}\int_{0}^{t}\mathsf{P}\bigg{\{}\sum_{i=1}^{n}X_{i}\leqslant x\,\bigg{|}\,\sum_{i=1}^{n}T_{i}=t-z\,\bigg{\}} (3.5)
×fTn(tz)𝖯{Tn+1>z}dz,\displaystyle\times f_{T}^{*n}(t-z)\,\mathsf{P}\big{\{}T_{n+1}>z\big{\}}\,dz,

which is the total probability formula.

Steps  B and C  remain essentially the same.

Step  D:  application of refined CLT.  The integrand in (3.5) is ready for application of (see [31][33]) two–dimensional hybrid integro–local (for density) CLT. In more detail, in the proof of Theorem 3.1, we use Berry–Esseen estimate, in the proof of Theorem 3.2, we use Edgeworth’s expansion in this CLT, both with non–uniform remainder terms.

Steps  E and F  remain nearly the same. Technically, they are more complicated than steps E and F in the proof of (3.1) and (3.2), but use essentially the same analytical techniques.

3.3. Garbage and deficiency of basic technique

When studying altered /\mathcal{R\hskip-0.5pt/I}  sum SN^t{S}_{\widehat{N}_{t}}, the main point of basic technique consists in approximating SN^t{S}_{\widehat{N}_{t}} by the ordinary sum S[𝖤N^t]{S}_{\,[\mathsf{E}\!\widehat{N}_{t}]}, whose asymptotical normality is evident: the real number [𝖤N^t][\mathsf{E}\!\widehat{N}_{t}] is (see (3.1)) of the order t/𝖤Tt/\mathsf{E}\,T\to\infty, as tt\to\infty, and the summands are i.i.d.  In other words, although technically the basic technique applies Kolmogorov’s inequality for maximum of partial sums, its fundamental idea is to throw the difference 𝒢t=SN^tS[𝖤N^t]\mathcal{G}_{\,t}={S}_{\widehat{N}_{t}}-{S}_{\,[\mathsf{E}\!\widehat{N}_{t}]}, called garbage, in the trash.

Constrained by this idea, the basic technique does not allow further refinements, such as in Theorems 3.1 and 3.2, since 𝒢t\mathcal{G}_{\,t} is a random variable of order t1/4t^{1/4}: under natural regularity conditions, including 𝖤(T3)<\mathsf{E}(T^{3})<\infty, 𝖤(X3)<\mathsf{E}(X^{3})<\infty,

supx𝖱|𝖯{μT3/4σXσT1/2t1/4𝒢tx}\displaystyle\sup_{x\in\mathsf{R}}\bigg{|}\,\mathsf{P}\bigg{\{}\frac{\mu_{T}^{3/4}}{\sigma_{X}\sigma_{T}^{1/2}t^{1/4}}\,\mathcal{G}_{\,t}\leqslant x\bigg{\}} (3.6)
Φ(0,1)(x|z|)φ(0,1)(z)dz|=O(t1/4),t.\displaystyle\hskip 40.0pt-\int_{-\infty}^{\infty}\varPhi_{\left({0},{1}\right)}\bigg{(}\frac{x}{\sqrt{|z|}}\bigg{)}\varphi_{\left({0},{1}\right)}(z)\,dz\bigg{|}=O\big{(}t^{-1/4}\big{)},\quad t\to\infty.

The proof of (3.6) (see details in [57] and [58]) by means of advanced technique follows the scheme sketched in the proof of (3.1) and (3.2) and Theorems 3.1 and 3.2. It consists of steps A–F, applies refined CLT with non–uniform remainder terms, and requires mere standard classical analysis at steps B, C, E, and F.

Refer to caption
Figure 2. Graph of f(y)=φ(0,1)(y|z|)1|z|φ(0,1)(z)𝑑zf(y)={\displaystyle\int_{-\infty}^{\infty}}\varphi_{\left({0},{1}\right)}\Big{(}\frac{y}{\sqrt{|z|}}\Big{)}\frac{1}{\sqrt{|z|}}\,\varphi_{\left({0},{1}\right)}(z)\,dz.

For greater clarity, note that p.d.f.  of the limit distribution in (3.6) is

f(y)=φ(0,1)(y|z|)1|z|φ(0,1)(z)𝑑z,f(y)=\int_{-\infty}^{\infty}\varphi_{\left({0},{1}\right)}\bigg{(}\frac{y}{\sqrt{|z|}}\bigg{)}\frac{1}{\sqrt{|z|}}\;\varphi_{\left({0},{1}\right)}(z)\,dz,

its maximal value is

f(0)=12π1|z|φ(0,1)(z)𝑑z=Γ(1/4)23/4π0.686,f(0)=\frac{1}{\sqrt{2\pi}}\int_{-\infty}^{\infty}\frac{1}{\sqrt{|z|}}\;\varphi_{\left({0},{1}\right)}(z)\,dz=\frac{\Gamma(1/4)}{2^{3/4}\pi}\approx 0.686,

and its graph is drawn in Fig. 2.

4. Limit theorems for 𝓘/𝓘\boldsymbol{\mathcal{I\hskip-0.5pt/I}}  sums and collective risk theory

The reader familiar with collective risk theory readily recognizes in 𝖯{SNtx}\mathsf{P}\{{S}_{N_{t}}\leqslant x\} of Example 2.1 the probability of ruin within time xx, where141414Under the assumption of independence from each other. XiX_{i}, i=1,2,i=1,2,\dots, are intervals between claims, YiY_{i}, i=1,2,i=1,2,\dots, are claim amounts, tt is the initial capital, and cc is the premium intensity. For 0<cc0<c\leqslant c^{\ast}, the /\mathcal{I\hskip-0.5pt/I}  sum SNt{S}_{N_{t}} is proper, while (in terms of the risk reserve model) insurance is ruinous. For c>c=𝖤Y/𝖤Tc>c^{\ast}=\mathsf{E}{Y}/\mathsf{E}{T}, the /\mathcal{I\hskip-0.5pt/I}  sum SNt{S}_{N_{t}} is defective, while insurance is profitable.

4.1. Modular approach: a way to manage complexity

A system is modular if its components can be separated and recombined, with the advantage of flexibility and variety in use. Breaking such a system into independent or nearly independent modules is done in order (see [10]) “to hide the complexity of each part behind an abstraction and interface”. Modularity is implemented in two ways: by splitting an orderly queue (for “queuing–type” models) by events that occur one after another in time, or by sorting spatial items (for “clustering–type” models) into loosely related and similar sets or groups.

An example of the latter is the development of the OS/360 operating system for the original IBM 360 line of computers. At first (see [16]), it was organized in a relatively indecomposable way. It was deemed that each programmer should be familiar with all the material, i.e., should have a copy of the workbook in his office. But over time, this led to a rapid growth of troubles. For example, even maintaining the workbook, whose volume constantly and significantly increased, began to take up significant time from each working day. Therefore, it became necessary to figure out whether programmers developing one particular module really need to know about the structure of the entire system.

It was concluded that, especially in large projects, the answer is negative and managers (quoting from [79]) “should pay attention to minimizing interdependencies. If knowledge is hidden or encapsulated within a module, that knowledge cannot affect, and therefore need not be communicated to, other parts of a system.” This idea has been framed as “information hiding”, a key concept in the modern object–oriented approach to computer programming.

While the modular approach is not needed for /𝒟\mathcal{R\hskip-0.5pt/D}  sums, it is useful when studying /\mathcal{I\hskip-0.5pt/I}, /𝒟\mathcal{R\hskip-0.5pt/D}, and proper /𝒟\mathcal{I\hskip-0.5pt/D}  sums. In “queuing–type” models, it is applied by splitting an orderly queue by events (see, e.g., [78]: “regenerative phenomena”, or the like) that occur one after another in time. In “clustering–type” models, it is applied by sorting spatial items into loosely related (and similar) sets or groups.

4.2. Normal approximation for proper 𝓘/𝓘\boldsymbol{\mathcal{I\hskip-0.5pt/I}}  sums

In this case, modular structure is defined by the random indices 00<1<20\equiv\mathbb{\ell}_{0}<\mathbb{\ell}_{1}<\mathbb{\ell}_{2}\dots, where

k=inf{i>k1:Vi>Vk1}=𝑑,k=1,2,.\mathbb{\ell}_{k}=\inf\big{\{}i>\mathbb{\ell}_{k-1}:{V}_{i}>{V}_{\mathbb{\ell}_{k-1}}\big{\}}\overset{d}{=}\mathbb{\ell},\quad k=1,2,\dots.

The key point is the following “exact”151515This means that, unlike, e.g., Dœblin’s dissection, it does not contain incomplete initial and final modules. partition of the proper /\mathcal{I\hskip-0.5pt/I}  sum SNt{S}_{N_{t}}, called Blackwell’s dissection161616This dissection was introduced by Blackwell in [13].:

SNt=i=1𝒩t𝒳i,{S}_{N_{t}}=\sum_{i=1}^{\mathcal{N}_{t}}{\mathcal{X}}_{\hskip 1.0pti}, (4.1)

where 𝒩t=inf{n>0:i=1n𝒯i>t}\mathcal{N}_{t}=\inf\big{\{}n>0:\sum_{i=1}^{n}{\mathcal{T}}_{\hskip 1.0pti}>t\big{\}}, or \infty, if the set is empty, 𝒳i=j=i1+1iXj{\mathcal{X}}_{\hskip 1.0pti}=\sum_{j=\mathbb{\ell}_{i-1}+1}^{\mathbb{\ell}_{i}}X_{j}, and 𝒯i=j=i1+1iTj>0{\mathcal{T}}_{\hskip 1.0pti}=\sum_{j=\mathbb{\ell}_{i-1}+1}^{\mathbb{\ell}_{i}}T_{j}>0 by definition. Bearing in mind that tt can be exceeded only at a ladder index, equality (4.1) is obvious.

Since (𝒯i,𝒳i)=𝑑(𝒯,𝒳)({\mathcal{T}}_{\hskip 1.0pti},{\mathcal{X}}_{\hskip 1.0pti})\overset{d}{=}({\mathcal{T}}_{\hskip 1.0pt},{\mathcal{X}}_{\hskip 1.0pt}), i=1,2,i=1,2,\dots, are i.i.d.  and 𝒯{\mathcal{T}}_{\hskip 1.0pt} is positive by definition, the sum on the right–hand side of (4.1) is a modular /\mathcal{R\hskip-0.5pt/I}  sum which is examined as above. Using (4.1), nearly all results on asymptotic normality and its refinements available for /\mathcal{R\hskip-0.5pt/I}  sums can be transferred to proper /\mathcal{I\hskip-0.5pt/I}  sums.

In particular (see details in [65], Chapter 9), in Example 2.1 with 0<c<c0<c<c^{\ast} the normal approximation under natural regularity conditions is

limtsupx𝖱|𝖯{SNtx}Φ(mt,D2t)(x)|=0,\lim_{t\to\infty}\sup_{x\in\mathsf{R}}\,\big{|}\,\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}}-\varPhi_{\left({m_{{}_{\scriptscriptstyle\triangledown}}t},{D_{{\hskip-1.0pt\scriptscriptstyle\triangledown}}^{2}t}\right)}(x)\,\big{|}=0, (4.2)

where m=μXμT1m_{{}_{\scriptscriptstyle\triangledown}}=\mu_{X}\mu_{T}^{-1} and D2=𝖤(μTXμXT)2μT3D_{{\hskip-1.0pt\scriptscriptstyle\triangledown}}^{2}=\mathsf{E}(\mu_{T}X-\mu_{X}T\,)^{2}\mu_{T}^{-3}. Its refinements, i.e., Berry–Esseen’s estimate and Edgeworth’s expansion, are similar to those in Theorems 3.1 and 3.2 and are omitted in this presentation.

Regarding mm_{{}_{\scriptscriptstyle\triangledown}} and D2D_{{\hskip-1.0pt\scriptscriptstyle\triangledown}}^{2} expressed above in the original (rather than modular) terms, the following should be noted. The refined normal approximation (4.2) obtained by modular approach is first written in terms of the modular random variables \mathbb{\ell}, 𝒯{\mathcal{T}}_{\hskip 1.0pt}, and 𝒳{\mathcal{X}}_{\hskip 1.0pt}. Converting them to the original terms requires additional effort. There are several ways to do this. First, Wald’s identities can bring a great release since certain combinations of moments of block random variables 𝒯{\mathcal{T}}_{\hskip 1.0pt} and 𝒳{\mathcal{X}}_{\hskip 1.0pt} can be represented through the moments of the original random variables TT and XX. Second, moments of ladder index \mathbb{\ell} and ladder modules 𝒯{\mathcal{T}}_{\hskip 1.0pt}, 𝒳{\mathcal{X}}_{\hskip 1.0pt} can be calculated analytically, using (see, e.g., [37]) Spitzer’s sums.

4.3. Quasi–normal approximation for defective 𝓘/𝓘\boldsymbol{\mathcal{I\hskip-0.5pt/I}}  sums

For c>cc>c^{\ast}, under natural regularity conditions (see details in [65], Chapter 9), quasi–normal approximation171717This name emphasizes the presence of a normal distribution function. In risk theory, it is known as Cramér–Lundberg’s approximation. is

limtsupx>0|eϰt𝖯{SNtx}Φ(mt,D2t)(x)|=0.\lim_{t\to\infty}\sup_{x>0}\,\big{|}\,e^{\varkappa t}\,\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}}-\mathbb{C}\;\varPhi_{\left({m_{{}_{\scriptscriptstyle\vartriangle}}t},{D_{{\hskip-1.0pt\scriptscriptstyle\vartriangle}}^{2}t}\right)}(x)\,\big{|}=0. (4.3)

It follows from (4.3) that 𝖯{SNt<}eϰt\mathsf{P}\{{S}_{N_{t}}<\infty\}\approx\,\mathbb{C}\,e^{-\varkappa t}, which is an asymptotic formula for defect (cf. (2.3)) of defective /\mathcal{I\hskip-0.5pt/I}  sum SNt{S}_{N_{t}}.

Here ϰ\varkappa is a positive solution (w.r.t.  xx) to the nonlinear equation181818In risk theory, equation (4.4) is called Lundberg’s equation. Its positive solution ϰ\varkappa is called Lundberg’s exponent, or adjustment coefficient.

𝖤exT=1\mathsf{E}{\,e^{xT}\,}=1 (4.4)

and in terms of the associated random variables T¯i=𝑑T¯\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu_{i}\overset{d}{=}\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu and X¯i=𝑑X¯\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu_{i}\overset{d}{=}\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu, i=1,2,i=1,2,\dots,

m=μX¯μT¯1,D2=𝖤(μT¯X¯μX¯T¯)2μT¯3,\displaystyle m_{{}_{\scriptscriptstyle\vartriangle}}=\mu_{\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu}\mu_{\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu}^{-1},\quad D_{{\hskip-1.0pt\scriptscriptstyle\vartriangle}}^{2}=\mathsf{E}\big{(}\mu_{\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu}\,\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu-\mu_{\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu}\,\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu\big{)}^{2}\mu_{\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu}^{-3},
=1ϰμT¯exp{n=11n𝖯{Vn>0}n=11n𝖯{V¯n0}},\displaystyle\mathbb{C}=\frac{1}{\varkappa\,\mu_{\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu}}\exp\bigg{\{}-\sum_{n=1}^{\infty}\frac{1}{n}\,\mathsf{P}\big{\{}{V}_{n}>0\big{\}}-\sum_{n=1}^{\infty}\frac{1}{n}\,\mathsf{P}\big{\{}{\mskip 1.5mu\overline{\mskip-2.0mu{V}\mskip-2.40002mu}\mskip 1.5mu}_{n}\leqslant 0\big{\}}\bigg{\}},

where V¯n=i=1nT¯i{\mskip 1.5mu\overline{\mskip-2.0mu{V}\mskip-2.40002mu}\mskip 1.5mu}_{n}=\sum_{i=1}^{n}\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu_{i}. Recall (see, e.g., Example (b) in [37], Chapter XII, Section 4) that the associated random variables we define as follows. Starting with /\mathcal{I\hskip-0.5pt/I}  basis (Ti,Xi)=𝑑(T,X)(T_{i},X_{i})\overset{d}{=}(T,X), i=1,2,i=1,2,\dots, we switch from FTX(t,x)=𝖯{Tt,Xx}F_{TX}(t,x)=\mathsf{P}\big{\{}T\leqslant t,X\leqslant x\big{\}} to FT¯X¯(t,x)=𝖯{T¯t,X¯x}F_{\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu}(t,x)=\mathsf{P}\big{\{}\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu\leqslant t,\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu\leqslant x\big{\}} defined by the integral cxt0xeϰuFTX(du,dv)\int_{-cx}^{t}\int_{0}^{x}e^{\varkappa u}\,F_{TX}(du,dv). The latter probability distribution is proper and 𝖤T¯>0\mathsf{E}\,\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu>0. Commonly used shorthand for it is FT¯X¯(dt,dx)=eϰzFTX(dt,dx)F_{\mskip 1.5mu\overline{\mskip-2.0mu{T}\mskip-1.20001mu}\mskip 0.90001mu\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu}(dt,dx)=e^{\varkappa z}F_{TX}(dt,dx).

Remark 4.1.

The proof of (4.3) carried out using the modular approach is based (see [9] and [63]) on the following counterpart of equality (4.1):

𝖯{SNtx}=𝖤(exp{ϰi=1𝒩¯t𝒯¯i} 1(,x](i=1𝒩¯t𝒳¯i)),\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\}=\mathsf{E}\bigg{(}\exp\bigg{\{}-\varkappa\;\sum_{i=1}^{{\bar{\mathcal{N}}}_{t}}\,{\bar{\mathcal{T}}}_{i}\bigg{\}}\,\boldsymbol{\mathsf{1}}_{(-\infty,x]}\bigg{(}\sum_{i=1}^{{\bar{\mathcal{N}}}_{t}}\,{\bar{\mathcal{X}}}_{i}\bigg{)}\bigg{)}, (4.5)

where 𝒩¯t=inf{n1:i=1n𝒯¯i>t}{\bar{\mathcal{N}}}_{t}=\inf\big{\{}n\geqslant 1:\sum_{i=1}^{n}{{\bar{\mathcal{T}}}_{i}}>t\big{\}}, or \infty, if the set is empty, and the simple modular basis (𝒯¯k,𝒳¯k)({\bar{\mathcal{T}}}_{k},{\bar{\mathcal{X}}}_{k}), k=1,2,k=1,2,\dots, is given by

𝒯¯k=V¯¯kV¯¯k1>0and𝒳¯k=S¯¯kS¯¯k1,k=1,2,,{\bar{\mathcal{T}}}_{k}={\mskip 1.5mu\overline{\mskip-2.0mu{V}\mskip-2.40002mu}\mskip 1.5mu}_{\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k}}-{\mskip 1.5mu\overline{\mskip-2.0mu{V}\mskip-2.40002mu}\mskip 1.5mu}_{\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k-1}}>0\quad\text{and}\quad{\bar{\mathcal{X}}}_{k}={\mskip 1.5mu\overline{\mskip-2.0mu{S}\mskip-2.40002mu}\mskip 1.5mu}_{\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k}}-{\mskip 1.5mu\overline{\mskip-2.0mu{S}\mskip-2.40002mu}\mskip 1.5mu}_{\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k-1}},\quad k=1,2,\dots,

where S¯n=i=1nX¯i{\mskip 1.5mu\overline{\mskip-2.0mu{S}\mskip-2.40002mu}\mskip 1.5mu}_{n}=\sum_{i=1}^{n}\mskip 3.59999mu\overline{\mskip-4.79999mu{X}\mskip-2.0mu}\mskip 1.5mu_{i} and the ladder indices generated by the associated random variables are ¯00\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{0}\equiv 0 and ¯k=inf{i>¯k1:V¯i>V¯¯k1}\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k}=\inf\big{\{}i>\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k-1}:{\mskip 1.5mu\overline{\mskip-2.0mu{V}\mskip-2.40002mu}\mskip 1.5mu}_{i}>{\mskip 1.5mu\overline{\mskip-2.0mu{V}\mskip-2.40002mu}\mskip 1.5mu}_{\,\mskip 0.59999mu\overline{\mskip-1.20001mu{\mathbb{\ell}}\mskip 1.20001mu}\mskip 1.5mu_{k-1}}\big{\}}, k=1,2,k=1,2,\dots.∎

Remark 4.2.

Edgeworth’s expansions are known for both the normal approximation (4.2) and (see [63]) the quasi–normal approximation (4.3). Further details see, e.g., in [65], Chapter 9.∎

Example 4.1 (Example 2.1 continued).

Let the random variables Xi=𝑑XX_{i}\overset{d}{=}X and Yi=𝑑YY_{i}\overset{d}{=}Y, i=1,2,i=1,2,\dots, be exponentially distributed with parameters ϱ>0\varrho>0 and ρ>0\rho>0. Clearly, 𝖤excX=ϱ/(ϱ+cx)\mathsf{E}{\,e^{-xcX}}=\varrho/(\varrho+cx) and 𝖤exY=ρ/(ρx)\mathsf{E}{\,e^{xY}}=\rho/(\rho-x). It follows that 𝖤exT=𝖤exY𝖤excX\mathsf{E}{\,e^{xT}\,}=\mathsf{E}{\,e^{xY}}\mathsf{E}{\,e^{-xcX}}, whence 𝖤exT=ρϱ/((ρx)(ϱ+cx))\mathsf{E}{\,e^{xT}\,}=\rho\varrho/((\rho-x)(\varrho+cx)), and equation (4.4) is the quadratic equation ρϱ/((ρx)(ϱ+cx))=1\rho\varrho/((\rho-x)(\varrho+cx))=1 with respect to xx. For c>c=ϱ/ρc>c^{\ast}=\varrho/\rho, its positive solution is ϰ=ρ(1ϱ/(cρ))\varkappa=\rho\,(1-\varrho/(c\rho)). Further calculations yield =ϱ/(cρ)\mathbb{C}=\varrho/(c\rho) and

m=1c(1ϱ/(cρ)),D2=2(ϱ/(cρ))c2ρ(1ϱ/(cρ))3,m=ϱ/(cρ)c(1ϱ/(cρ)),D2=2(ϱ/(cρ))c2ρ(1ϱ/(cρ))3.\begin{gathered}m_{{}_{\scriptscriptstyle\triangledown}}=-\frac{1}{c\,(1-\varrho/(c\rho))},\quad D_{{\hskip-1.0pt\scriptscriptstyle\triangledown}}^{2}=-\frac{2\,(\varrho/(c\rho))}{c^{2}\rho\,(1-\varrho/(c\rho))^{3}},\\[0.0pt] m_{{}_{\scriptscriptstyle\vartriangle}}=\frac{\varrho/(c\rho)}{c\,(1-\varrho/(c\rho))},\quad D_{{\hskip-1.0pt\scriptscriptstyle\vartriangle}}^{2}=\frac{2\,(\varrho/(c\rho))}{c^{2}\rho\,(1-\varrho/(c\rho))^{3}}.\end{gathered}
Refer to caption
Figure 3. Graph (XX-axis is cc) of 𝖯{SNtx}\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}} (solid line), which is a monotone decreasing function calculated numerically, using equalities (2.2)–(2.4), and the approximation (4.2), (4.3) (dash–dotted line), when TT and YY are exponentially distributed with parameters δ=2\delta=2 and ρ=1\rho=1, respectively, and t=200t=200, u=10u=10. Horizontal line: 𝖯{SNtx}\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}} with cc equal to cc^{\ast}, which is 0.6990.699.
Refer to caption
Figure 4. Graphs (XX-axis is cc) of 𝖯{SNtx}\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}} (solid line) calculated numerically, using equalities (2.2)–(2.4), and approximation (4.6) (dash–dotted line), when TT and YY are exponentially distributed with parameters δ=2\delta=2 and ρ=1\rho=1, respectively, and t=200t=200, u=10u=10. Horizontal line: 𝖯{SNtx}\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}} with cc equal to cc^{\ast}, which is 0.6990.699.

Approximations (4.2) and (4.3) are compared (see Fig. 3) with the exact values of 𝖯{SNtx}\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}} calculated numerically. They are unsatisfactory in the vicinity of the point cc^{\ast} and satisfactory elsewhere. ∎

Remark 4.3.

The assumption that a positive solution ϰ\varkappa to the equation (4.4) exists is a serious constraint on T=YcXT=Y-cX. It follows that 𝖤exT\mathsf{E}{\,e^{xT}\,} must exist in the right neighborhood of 0 and that 𝖯{T>x}\mathsf{P}\big{\{}T>x\big{\}} is exponentially bounded from above for x>0x>0. The latter follows from Markov’s inequality, as follows: 𝖯{T>x}eϰx𝖤eϰT=eϰx\mathsf{P}\big{\{}T>x\big{\}}\leqslant e^{-\varkappa x}\,\mathsf{E}{\,e^{\varkappa T}}=e^{-\varkappa x}.∎

4.4. Inverse Gaussian approximation for 𝓘/𝓘\boldsymbol{\mathcal{I\hskip-0.5pt/I}}  sums

The asymptotic (as tt\to\infty) behavior of /\mathcal{I\hskip-0.5pt/I}  sum does not necessarily have to be related to a normal distribution. The following inverse Gaussian approximation (see details in [65], [66]) for 𝖯{SNtx}\mathsf{P}\{{S}_{N_{t}}\leqslant x\} has great advantages. In the framework of Example 2.1, introduce M=𝖤X/𝖤YM=\mathsf{E}{X}/\mathsf{E}{Y} and D 2=((𝖤X)2𝖣Y+(𝖤Y)2𝖣X)/(𝖤Y)3D^{\,2}=((\mathsf{E}{X})^{2}\mathsf{D}{Y}+(\mathsf{E}{Y})^{2}\mathsf{D}{X})/(\mathsf{E}{Y})^{3}. Under certain regularity conditions, we have

limtsupx>0|𝖯{SNtx}t,c(x)|=0,\lim_{t\to\infty}\sup_{x>0}\big{|}\,\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}}-{\mathcal{M}}_{t,c}(x)\,\big{|}=0, (4.6)

where

t,c(x)\displaystyle{\mathcal{M}}_{t,c}(x) =0cxt1z+1φ(cM(z+1),c2D2u(z+1))(z)𝑑z,\displaystyle=\int_{0}^{\frac{cx}{t}}\frac{1}{z+1}\,\varphi_{\left({cM(z+1)},{\frac{c^{2}D^{2}}{u}(z+1)}\right)}(z)\,dz,

or, in equivalent form,

t,c(x)\displaystyle{\mathcal{M}}_{t,c}(x) ={(F(z+1;μ,λ)F(1;μ,λ))|z=cxt,μ=11cM,λ=tc2D2,0<cc,exp{2λμ^}(F(z+1;μ^,λ)F(1;μ^,λ))|z=cxt,μ^=1cM1,λ=tc2D2,c>c\displaystyle=\begin{cases}\big{(}F(z+1;\mu,\lambda)\\[2.0pt] \hskip 30.0pt-F(1;\mu,\lambda)\big{)}\Big{|}_{z=\frac{cx}{t},\mu=\frac{1}{1-cM},\lambda=\frac{t}{c^{2}D^{2}}},&0<c\leqslant c^{\ast},\\[2.0pt] \exp\bigg{\{}-\dfrac{2\lambda}{\hat{\mu}}\bigg{\}}\,\big{(}F(z+1;\hat{\mu},\lambda)\\[6.0pt] \hskip 30.0pt-F(1;\hat{\mu},\lambda)\big{)}\Big{|}_{z=\frac{cx}{t},\hat{\mu}=\frac{1}{cM-1},\lambda=\frac{t}{c^{2}D^{2}}},&c>c^{\ast}\end{cases}

and

F(z;μ,λ)=\displaystyle F(z;\mu,\lambda)= Φ(0,1)(λz(zμ1))\displaystyle\,\varPhi_{\left({0},{1}\right)}\bigg{(}\sqrt{\frac{\lambda}{z}}\,\bigg{(}\frac{z}{\mu}-1\bigg{)}\bigg{)} (4.7)
+exp{2λμ}Φ(0,1)(λz(zμ+1)),z>0.\displaystyle+\exp\bigg{\{}\frac{2\lambda}{\mu}\bigg{\}}\,\varPhi_{\left({0},{1}\right)}\bigg{(}-\sqrt{\frac{\lambda}{z}}\,\bigg{(}\frac{z}{\mu}+1\bigg{)}\bigg{)},\quad z>0.

The approximation (4.6) is called inverse Gaussian because (4.7) is c.d.f.  of the inverse Gaussian distribution with parameters μ>0\mu>0 and λ>0\lambda>0.

Remark 4.4.

In [65], Chapter 7, refinements (i.e., analogues of the Berry–Esseen estimate and the Edgeworth expansion) of the inverse Gaussian approximation (4.6) are obtained.∎

Comparing approximations (4.2) and (4.3) with (4.6), one can argue191919Using heuristic arguments and referring to exact formulas (2.2)–(2.4). that the results for proper and defective /\mathcal{I\hskip-0.5pt/I}  sums in Example 2.1 must border seamlessly. But the normal (with 0<c<c=𝖤Y/𝖤T0<c<c^{\ast}=\mathsf{E}{Y}/\mathsf{E}{T}) and quasi–normal (with c>cc>c^{\ast}) approximations (4.2) and (4.3) do not border seamlessly. Moreover (see Fig. 3), both of them are either poor, or invalid202020Quasi–normal approximation (4.3) formally requires c>cc>c^{\ast}. The existence of ϰ>0\varkappa>0 is possible only if the tail of YY is exponentially decreasing. for cc at or near the critical value cc^{\ast}.

This distressing fact only indicates that the normal and quasi–normal approximations are deficient at cc^{\ast}. In contrast, the inverse Gaussian approximation (see Fig. 4) does not have such a drawback. The case when cc is close to cc^{\ast} or even equals cc^{\ast} is routinely covered by the inverse Gaussian approximation.

Remark 4.5.

The intuitive explanation of deficiency of normal and quasi–normal approximations in the vicinity of the critical value cc^{\ast} is given in [64], Section 9.2.4. It is as follows. For exponentially distributed XiX_{i} and YiY_{i}, i=1,2,i=1,2,\dots, and cc equal to cc^{\ast} the compound sum is made up of the random variables with power moments of order no greater than 1/21/2. Therefore, there is no reason to expect validity of normal and quasi–normal approximations.∎

The flaw of (4.2) and (4.3) in the vicinity od cc^{\ast} is merely a drawback of a particular mathematical technique. However, it has had a major negative impact on insurance modeling: while the insurance process in a close neighborhood of the critical point cc^{\ast} is of great importance (see details in [64]), it was overlooked for a long time partly due to this flaw. Moreover, the inverse Gaussian approximation allows (see [66]) to develop a new approach to risk measures different from value at risk (VaR).

5. Limit theorems for 𝓡/𝓓\boldsymbol{\mathcal{R\hskip-0.5pt/D}} and proper 𝓘/𝓓\boldsymbol{\mathcal{I\hskip-0.5pt/D}}  sums

Kesten mentioned the part of modular analysis that deals with the lack of independence. It is best known because the area of “Markov chains, continuous time processes, and other situations in which the independence between summands no longer applies” is of particular interest. Among the tricks which reduce limit theorems for dependent summands to the case of independent summands, Kesten mentioned “Dœblin’s trick”.

Studying limit theorems for /𝒟\mathcal{R\hskip-0.5pt/D} and proper /𝒟\mathcal{I\hskip-0.5pt/D}  sums, we will combine “Dœblin’s trick” and “Blackwell’s trick” and extend it to compound sums with general modular structure.

5.1. Compound sums with general modular structure

For t>0t>0 and a complex basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, we focus on unaltered complex compound sum SNt{S}_{N_{t}}. We denote by 0<𝒥0<𝒥1<0<\mathcal{J}_{0}<\mathcal{J}_{1}<\dots a sequence of random indices, put

τ0=𝒥0,𝒯0=i=1𝒥0Ti,𝒳0=i=1𝒥0Xi,τk=𝒥k𝒥k1,𝒯k=i=𝒥k1+1𝒥kTi,𝒳k=i=𝒥k1+1𝒥kXi,\begin{gathered}{\mathcal{\tau}}_{\hskip 1.0pt0}=\mathcal{J}_{0},\quad{\mathcal{T}}_{\hskip 1.0pt0}=\sum_{i=1}^{\mathcal{J}_{0}}\;T_{i},\quad{\mathcal{X}}_{\hskip 1.0pt0}=\sum_{i=1}^{\mathcal{J}_{0}}\;X_{i},\\[0.0pt] {\mathcal{\tau}}_{\hskip 1.0ptk}=\mathcal{J}_{k}-\mathcal{J}_{k-1},\quad{\mathcal{T}}_{\hskip 1.0ptk}=\sum_{i=\mathcal{J}_{k-1}+1}^{\mathcal{J}_{k}}T_{i},\quad{\mathcal{X}}_{\hskip 1.0ptk}=\sum_{i=\mathcal{J}_{k-1}+1}^{\mathcal{J}_{k}}X_{i},\end{gathered}

where k=1,2,k=1,2,\dots, and introduce the following definition.

Definition 5.1.

We say that the basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, allows a modular structure if there exist a.s.  finite random indices 0<𝒥0<𝒥1<0<\mathcal{J}_{0}<\mathcal{J}_{1}<\dots, such that

  • (i)

    the random vectors (τk,𝒯k,𝒳k)({\mathcal{\tau}}_{\hskip 1.0ptk},{\mathcal{T}}_{\hskip 1.0ptk},{\mathcal{X}}_{\hskip 1.0ptk}) are independent for k=0,1,k=0,1,\dots, and identically distributed for k=1,2,k=1,2,\dots,

  • (ii)

    V𝒥k>max0i𝒥k1Vi{V}_{\mathcal{J}_{k}}>\max_{0\leqslant i\leqslant\mathcal{J}_{k}-1}{V}_{i}, k=0,1,k=0,1,\dots, i.e., 𝒥i\mathcal{J}_{i}, i=0,1,i=0,1,\dots, are upper record moments of the sequence Vn=i=1nTi{V}_{n}=\sum_{i=1}^{n}T_{i}, n=1,2,n=1,2,\dots.

For modular i.i.d.  random vectors (𝒳k,𝒯k,τk)=𝑑(𝒳,𝒯,τ)({\mathcal{X}}_{\hskip 1.0ptk},{\mathcal{T}}_{\hskip 1.0ptk},{\mathcal{\tau}}_{\hskip 1.0ptk})\overset{d}{=}({\mathcal{X}}_{\hskip 1.0pt},{\mathcal{T}}_{\hskip 1.0pt},{\mathcal{\tau}}_{\hskip 1.0pt}), k=1,2,k=1,2,\dots, we use the shorthand notation μτ=𝖤τ\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}=\mathsf{E}{\mathcal{\tau}}_{\hskip 1.0pt}, μ𝒯=𝖤𝒯\mu_{{\mathcal{T}}_{\hskip 1.0pt}}=\mathsf{E}{\mathcal{T}}_{\hskip 1.0pt}, μ𝒳=𝖤𝒳\mu_{{\mathcal{X}}_{\hskip 1.0pt}}=\mathsf{E}{\mathcal{X}}_{\hskip 1.0pt}, στ2=𝖣τ\sigma_{{\mathcal{\tau}}_{\hskip 1.0pt}}^{2}=\mathsf{D}{\mathcal{\tau}}_{\hskip 1.0pt}, σ𝒯2=𝖣𝒯\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}^{2}=\mathsf{D}{\mathcal{T}}_{\hskip 1.0pt}, σ𝒳2=𝖣𝒳\sigma_{{\mathcal{X}}_{\hskip 1.0pt}}^{2}=\mathsf{D}{\mathcal{X}}_{\hskip 1.0pt}. For τ^=τμτ\hat{\mathcal{\tau}}^{\hskip 1.0pt}={\mathcal{\tau}}_{\hskip 1.0pt}-\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}, 𝒯^=𝒯μ𝒯\hat{\mathcal{T}}^{\hskip 1.0pt}={\mathcal{T}}_{\hskip 1.0pt}-\mu_{{\mathcal{T}}_{\hskip 1.0pt}}, 𝒳^=𝒳μ𝒳\hat{\mathcal{X}}^{\hskip 1.0pt}={\mathcal{X}}_{\hskip 1.0pt}-\mu_{{\mathcal{X}}_{\hskip 1.0pt}}, and write ϰ𝒯τ=𝖤𝒯^τ^\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}=\mathsf{E}\hat{\mathcal{T}}^{\hskip 1.0pt}\!\hat{\mathcal{\tau}}^{\hskip 1.0pt}, ϰ𝒳𝒯=𝖤𝒳^𝒯^\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}=\mathsf{E}\hat{\mathcal{X}}^{\hskip 1.0pt}\hat{\mathcal{T}}^{\hskip 1.0pt}, ϰ𝒳τ=𝖤𝒳^τ^\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}=\mathsf{E}\hat{\mathcal{X}}^{\hskip 1.0pt}\hat{\mathcal{\tau}}^{\hskip 1.0pt}, ρ𝒯τ=ϰ𝒯τ/(σ𝒯στ)\rho_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}=\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}/(\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}\sigma_{{\mathcal{\tau}}_{\hskip 1.0pt}}) ρ𝒳𝒯=ϰ𝒳𝒯/(σ𝒳σ𝒯)\rho_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}=\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}/(\sigma_{{\mathcal{X}}_{\hskip 1.0pt}}\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}) ρ𝒳τ=ϰ𝒳τ/(σ𝒳στ)\rho_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}=\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}/(\sigma_{{\mathcal{X}}_{\hskip 1.0pt}}\sigma_{{\mathcal{\tau}}_{\hskip 1.0pt}}). We further introduce zero–mean block random variables212121Note that 𝖤(missing𝒳𝒯2)\mathsf{E}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{X}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}^{2}\big{)} is the same as 𝖣(missing𝒳𝒯)\mathsf{D}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{X}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}\big{)} and 𝖤(missingτ𝒯2)\mathsf{E}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{\tau}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}^{2}\big{)} is the same as 𝖣(missingτ𝒯)\mathsf{D}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{\tau}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}\big{)}.

missing𝒯τ\displaystyle{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}} =μ𝒯τμτ𝒯\displaystyle=\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\,{\mathcal{\tau}}_{\hskip 1.0pt}-\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}{\mathcal{T}}_{\hskip 1.0pt} missing𝒳𝒯=μ𝒯𝒳μ𝒳𝒯\displaystyle{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{X}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}=\mu_{{\mathcal{T}}_{\hskip 1.0pt}}{\mathcal{X}}_{\hskip 1.0pt}-\mu_{{\mathcal{X}}_{\hskip 1.0pt}}{\mathcal{T}}_{\hskip 1.0pt}
   and
=μ𝒯τ^μτ𝒯^,\displaystyle=\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\hat{\mathcal{\tau}}^{\hskip 1.0pt}-\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}\hat{\mathcal{T}}^{\hskip 1.0pt}, =μ𝒯𝒳^μ𝒳𝒯^\displaystyle=\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\hat{\mathcal{X}}^{\hskip 1.0pt}-\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\hat{\mathcal{T}}^{\hskip 1.0pt}

and put

N=μτμ𝒯1,𝒟N 2=𝖤(missing𝒯τ2)μ𝒯3,S=μ𝒳μ𝒯1,𝒟S 2=𝖤(missing𝒳𝒯2)μ𝒯3,SN=𝖤(missing𝒯τmissing𝒳𝒯)μ𝒯3,ρSN=SN(𝒟S𝒟N)1.\begin{gathered}\mathcal{M}_{N}{\,\!}=\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-1},\quad\mathcal{D}_{N}^{\,2}{\,}=\mathsf{E}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}^{2}\big{)}\,\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-3},\\[2.0pt] \mathcal{M}_{{S}}{\hskip 0.6pt\!}=\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-1},\quad\mathcal{D}_{{S}}^{\,2}{\,}=\mathsf{E}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{X}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}^{2}\big{)}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-3},\\[0.0pt] \mathcal{B}_{{S}N}{\,\!}=\mathsf{E}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{X}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}\big{)}\,\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-3},\quad\mathcal{\rho}_{{S}\!N}{\,\!}=\mathcal{B}_{{S}N}{\,}\big{(}\mathcal{D}_{{S}}^{\,}{\,}\mathcal{D}_{N}^{\,}{\,\!}\big{)}^{-1}.\end{gathered}

Assumption (i) tackles “dependence complexity”, while (ii) takes hold of “irregularity complexity”, in both cases by reducing the original complex compound sum to a simple modular compound sum with simple modular basis (𝒯k,𝒳k)({\mathcal{T}}_{\hskip 1.0ptk},{\mathcal{X}}_{\hskip 1.0ptk}), k=1,2,k=1,2,\dots. It consists of i.i.d.  modular random vectors with ss–components positive. Moreover, if V𝒥k<t{V}_{\mathcal{J}_{k}}<t, then crossing level tt cannot occur inside kkth block or inside blocks with a smaller index.

With 𝖱(x,y)=xρSNy1ρSN2\mathsf{R}_{\,}(x,y)=\dfrac{x-\mathcal{\rho}_{{S}\!N}{\,\!}y}{1-\mathcal{\rho}_{{S}\!N}^{2}{\,}}, we will use the following notation:

𝒦(3,0)=𝖤(missing𝒳𝒯3)μ𝒯4𝒟S33𝖱(L1,L2),\mathcal{K}_{(3,0)}=\,\mathsf{E}\big{(}{\mathcal{R_{missing}}{}}_{\hskip 1.0pt\!{\mathcal{X}}_{\hskip 1.0pt}\!{\mathcal{T}}_{\hskip 1.0pt}}^{3}\big{)}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-4}\mathcal{D}_{{S}}^{\,-3}{\,}-3\mathsf{R}_{\,}({L}_{1},{L}_{2}),

where

L1=\displaystyle{L}_{1}= (ϰ𝒳τ(μτσ𝒯2μ𝒯ϰ𝒯τ)στ2(μ𝒳σ𝒯2μ𝒯ϰ𝒳𝒯)\displaystyle\,\big{(}\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}(\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}^{2}-\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}})-\sigma_{{\mathcal{\tau}}_{\hskip 1.0pt}}^{2}(\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}^{2}-\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}})
+ϰ𝒯τ(μ𝒳ϰ𝒯τμτϰ𝒳𝒯))μ𝒯3𝒟S1𝒟N2,\displaystyle+\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}(\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}-\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}})\big{)}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-3}\mathcal{D}_{{S}}^{\,-1}{\,}\mathcal{D}_{N}^{\,-2}{\,},
L2=\displaystyle{L}_{2}= (ϰ𝒳τ(μ𝒳σ𝒯2μ𝒯ϰ𝒳𝒯)+σ𝒳2(μτσ𝒯2μ𝒯ϰ𝒯τ)\displaystyle\,\big{(}-\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}(\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}^{2}-\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}})+\sigma_{{\mathcal{X}}_{\hskip 1.0pt}}^{2}(\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}^{2}-\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}})
ϰ𝒳𝒯(μτϰ𝒳𝒯μ𝒳ϰ𝒯τ))μ𝒯3𝒟S2𝒟N1.\displaystyle-\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}(\mu_{{\mathcal{\tau}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}-\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}})\big{)}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-3}\mathcal{D}_{{S}}^{\,-2}{\,}\mathcal{D}_{N}^{\,-1}{\,}.

It can be shown by direct algebra that 𝖱(L1,L2)=(μ𝒳σ𝒯2μ𝒯ϰ𝒳𝒯)𝒟S1μ𝒯2\mathsf{R}_{\,}\big{(}{L}_{1},{L}_{2}\big{)}=-\big{(}\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\sigma_{{\mathcal{T}}_{\hskip 1.0pt}}^{2}-\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\varkappa_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}\big{)}\,\mathcal{D}_{{S}}^{\,-1}{\,}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-2}. Finally, we introduce

η1\displaystyle\eta_{\hskip 1.0pt1} =𝖤𝒳0+μ𝒯1s=1x0𝖯{τ1s,𝒩(t)=s,𝒳0(s)dx}𝑑t,\displaystyle=\mathsf{E}{\mathcal{X}}_{\hskip 1.0pt0}+\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-1}\sum_{s=1}^{\infty}\int_{-\infty}^{\infty}x\int_{0}^{\infty}\mathsf{P}\big{\{}{\mathcal{\tau}}_{\hskip 1.0pt1}\geqslant s,{\mathcal{N}}_{\hskip 1.0pt}(t)=s,{\mathcal{X}}_{\hskip 1.0pt0}^{\hskip 1.0pt(s)}\in dx\big{\}}\,dt,
η2\displaystyle\eta_{\hskip 1.0pt2} =𝖤𝒯0+μ𝒯1s=10t𝖯{τ1s,𝒩(t)=s}𝑑t,\displaystyle=\mathsf{E}{\mathcal{T}}_{\hskip 1.0pt0}+\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-1}\sum_{s=1}^{\infty}\int_{0}^{\infty}t\,\mathsf{P}\big{\{}{\mathcal{\tau}}_{\hskip 1.0pt1}\geqslant s,{\mathcal{N}}_{\hskip 1.0pt}(t)=s\big{\}}\,dt,
η3\displaystyle\eta_{\hskip 1.0pt3} =𝖤τ0+μ𝒯1s=1s0𝖯{τ1s,𝒩(t)=s}𝑑t,\displaystyle=\mathsf{E}{\mathcal{\tau}}_{\hskip 1.0pt0}+\mu_{{\mathcal{T}}_{\hskip 1.0pt}}^{-1}\sum_{s=1}^{\infty}s\int_{0}^{\infty}\mathsf{P}\big{\{}{\mathcal{\tau}}_{\hskip 1.0pt1}\geqslant s,{\mathcal{N}}_{\hskip 1.0pt}(t)=s\big{\}}\,dt,

where 𝒳0(s)=k=𝒥0+1𝒥0+sXk{\mathcal{X}}_{\hskip 1.0pt0}^{\hskip 1.0pt(s)}=\sum_{k=\mathcal{J}_{0}+1}^{\mathcal{J}_{0}+s}X_{k} and 𝒩(t)=inf{n1:k=𝒥0+1𝒥0+nTk>t}{\mathcal{N}}_{\hskip 1.0pt}(t)=\inf\big{\{}n\geqslant 1:\sum_{k=\mathcal{J}_{0}+1}^{\mathcal{J}_{0}+n}T_{k}>t\big{\}}, or \infty, if the set is empty.

Theorem 5.1 below is formulated under the following conditions.

Condition 𝑽\boldsymbol{V} (Non–degenerate correlation matrix) 𝖽𝖾𝗍Q>0\mathsf{det}\,{{{Q}}}>0 for the correlation matrix

Q=Cor(𝒳1,𝒯1,τ1)=(1ρ𝒳𝒯ρ𝒳τρ𝒳𝒯1ρ𝒯τρ𝒳τρ𝒯τ1).{{{Q}}}={Cor}\big{(}{\mathcal{X}}_{\hskip 1.0pt1},{\mathcal{T}}_{\hskip 1.0pt1},{\mathcal{\tau}}_{\hskip 1.0pt1}\big{)}=\left(\begin{matrix}1&\rho_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}&\rho_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}\\[0.0pt] \rho_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{T}}_{\hskip 1.0pt}}&1&\rho_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}\\[0.0pt] \rho_{{\mathcal{X}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}&\rho_{{\mathcal{T}}_{\hskip 1.0pt}{\mathcal{\tau}}_{\hskip 1.0pt}}&1\end{matrix}\right).

Condition 𝑷\boldsymbol{P}_{\boldsymbol{}} (Bounded density condition) For an integer n1n\geqslant 1, there exists a bounded convolution power f𝒯,τn(x,n)f_{{\mathcal{T}}_{\hskip 1.0pt},{\mathcal{\tau}}_{\hskip 1.0pt}}^{*n}(x,n) w.r.t.  Lebesgue measure.

Condition 𝑳\boldsymbol{L} (Lattice condition) The integer–valued random variable τ1{\mathcal{\tau}}_{\hskip 1.0pt1} assumes values in the set of natural numbers with maximal span 11.

Condition 𝑪\boldsymbol{C} (Uniform Cramér’s condition)

lim¯|t1||𝖤exp{it1𝒳+it2𝒯+it3τ}|<1\overline{\lim}_{|t_{1}|\to\infty}\big{|}\,\mathsf{E}\exp\big{\{}it_{1}{\mathcal{X}}_{\hskip 1.0pt}+it_{2}{\mathcal{T}}_{\hskip 1.0pt}+it_{3}{\mathcal{\tau}}_{\hskip 1.0pt}\big{\}}\big{|}<1

for all t2,t3𝖱t_{2},t_{3}\in\mathsf{R}.

Condition 𝑩(𝒓,𝒔)\boldsymbol{B}_{\boldsymbol{(r,s)}} (Block moments condition) For r,s>0r,s>0

  • (i)

    𝖤(𝒥0r)<\mathsf{E}\hskip 1.0pt\big{(}\mathcal{J}_{0}^{\hskip 1.0ptr}\big{)}<\infty, 𝖤(τ1s)<\mathsf{E}\big{(}{\mathcal{\tau}}_{\hskip 1.0pt1}^{s}\big{)}<\infty,

  • (ii)

    𝖤([k=1𝒥0|Tk|]r)<\mathsf{E}\bigg{(}\bigg{[}\displaystyle\sum_{k=1}^{\mathcal{J}_{0}}\big{|}\,T_{k}\big{|}\,\bigg{]}^{r}\bigg{)}<\infty, 𝖤([k=𝒥0+1𝒥1|Tk|]s)<\mathsf{E}\bigg{(}\bigg{[}\displaystyle\sum_{k=\mathcal{J}_{0}+1}^{\mathcal{J}_{1}}\big{|}\,T_{k}\big{|}\,\bigg{]}^{\hskip-1.0pts}\bigg{)}<\infty,

  • (iii)

    𝖤([k=1𝒥0|Xk|]r)<\mathsf{E}\bigg{(}\bigg{[}\displaystyle\sum_{k=1}^{\mathcal{J}_{0}}\big{|}\,X_{k}\big{|}\,\bigg{]}^{r}\bigg{)}<\infty, 𝖤([k=𝒥0+1𝒥1|Xk|]s)<\mathsf{E}\bigg{(}\bigg{[}\displaystyle\sum_{k=\mathcal{J}_{0}+1}^{\mathcal{J}_{1}}\big{|}\,X_{k}\big{|}\,\bigg{]}^{\hskip-1.0pts}\,\bigg{)}<\infty.

The following theorem is Theorem 1 (III) in [62].

Theorem 5.1.

Assume that the basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, of the complex compound sum SNt{S}_{N_{t}} allows a modular structure. If conditions V{V}, P{P}, L{L}, C{C}, and B(k32,k){B}_{(k-\frac{3}{2},k)} with k>3k>3 are satisfied, then there exist polynomials 𝖰r(x)\mathsf{Q}_{\,r}(x), r=1,2,,k3r=1,2,\dots,k-3, of degree 3r13r-1 such that

supx𝖱|𝖯{SNtStx𝒟St1/2}Φ(0,1)(x)\displaystyle\sup_{x\in\mathsf{R}}\,\bigg{|}\,\mathsf{P}\big{\{}{S}_{N_{t}}-\mathcal{M}_{{S}}{\hskip 0.6ptt}\leqslant x\mathcal{D}_{{S}}^{\,}{\,t^{1/2}}\big{\}}-\varPhi_{\left({0},{1}\right)}(x)
r=1k3𝖰r(x)tr/2φ(0,1)(x)|=O(t(k2)/2),t.\displaystyle\hskip 90.0pt-\sum_{r=1}^{k-3}\frac{\mathsf{Q}_{\,r}(x)}{t^{r/2}}\,\varphi_{\left({0},{1}\right)}(x)\,\bigg{|}=O\big{(}t^{(k-2)/2}\big{)},\quad t\to\infty.

In particular,

𝖰 1(x)=16𝒦(3,0)𝖧2(x)+𝖱(L1,L2)μ𝒯η1μ𝒳η2𝒟Sμ𝒯.\mathsf{Q}_{\,1}(x)=-\frac{1}{6}\;\mathcal{K}_{(3,0)}\mathsf{H}_{2}(x)+\mathsf{R}_{\,}({L}_{1},{L}_{2})-\frac{\mu_{{\mathcal{T}}_{\hskip 1.0pt}}\eta_{\hskip 1.0pt1}-\mu_{{\mathcal{X}}_{\hskip 1.0pt}}\eta_{\hskip 1.0pt2}}{\mathcal{D}_{{S}}^{\,}{\,}\mu_{{\mathcal{T}}_{\hskip 1.0pt}}}.
Sketch of proof of Theorem 5.1.

The proof, similarly to the proof of Theorem 3.2, consists of steps A–F.

Step  A:  use of fundamental identity.  Similar to (3.5), we use the total probability rule. Writing

𝒳[r]=i=1rXi,𝒯[r]=i=1rTi,𝒳(s)=i=𝒥m+1𝒥m+sXi,𝒯(s)=i=𝒥m+1𝒥m+sTi\begin{gathered}{\mathcal{X}}_{\hskip 1.0pt}^{[r]}=\sum_{i=1}^{r}X_{i},\ {\mathcal{T}}_{\hskip 1.0pt}^{[r]}=\sum_{i=1}^{r}T_{i},\ {\mathcal{X}}_{\hskip 1.0pt}^{(s)}=\sum_{i=\mathcal{J}_{m}+1}^{\mathcal{J}_{m}+s}X_{i},\ {\mathcal{T}}_{\hskip 1.0pt}^{(s)}=\sum_{i=\mathcal{J}_{m}+1}^{\mathcal{J}_{m}+s}T_{i}\end{gathered}

and 𝒩(t)=inf{s1:𝒯(s)>t}{\mathcal{N}}_{\hskip 1.0pt}(t)=\inf\big{\{}s\geqslant 1:{\mathcal{T}}_{\hskip 1.0pt}^{(s)}>t\big{\}}, we have

𝖯{SNtx}=n=1𝖯{SNtx,Nt=n},\displaystyle\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x\big{\}}=\,\sum_{n=1}^{\infty}\mathsf{P}\big{\{}{S}_{N_{t}}\leqslant x,N_{t}=n\big{\}},

where the right–hand side is the sum of, first, the expression

n=1r0t1t0x1x𝖯{τ0=r,𝒯[r]dt1,𝒳[r]dx1}\displaystyle\sum_{n=1}^{\infty}\sum_{r}\iint_{\begin{subarray}{c}0\leqslant t_{1}\leqslant t\\[1.0pt] 0\leqslant x_{1}\leqslant x\end{subarray}}\mathsf{P}\Big{\{}{\mathcal{\tau}}_{\hskip 1.0pt0}=r,{\mathcal{T}}_{\hskip 1.0pt}^{[r]}\in dt_{1},{\mathcal{X}}_{\hskip 1.0pt}^{[r]}\in dx_{1}\Big{\}}
×𝖯{τ1nr,𝒩(tt1)=nr,𝒳(nr)xx1},\displaystyle\hskip 30.0pt\times\mathsf{P}\Big{\{}{\mathcal{\tau}}_{\hskip 1.0pt1}\geqslant n-r,{\mathcal{N}}_{\hskip 1.0pt}(t-t_{1})=n-r,{\mathcal{X}}_{\hskip 1.0pt}^{(n-r)}\leqslant x-x_{1}\Big{\}},

which corresponds to the absence of complete blocks, and, second, the expression

n=1r,sm=1n(r+s)∫⋯∫0t1+t2t0x1+x2x𝖯{i=1m𝒳ix(x1+x2),\displaystyle\sum_{n=1}^{\infty}\sum_{r,s}\sum_{m=1}^{n-(r+s)}\idotsint_{\begin{subarray}{c}0\leqslant t_{1}+t_{2}\leqslant t\\[1.0pt] 0\leqslant x_{1}+x_{2}\leqslant x\end{subarray}}\mathsf{P}\bigg{\{}\sum_{i=1}^{m}{\mathcal{X}}_{\hskip 1.0pti}\leqslant x-\big{(}x_{1}+x_{2}\big{)}, (5.1)
i=1mτi=n(s+r)|i=1m𝒯i=t(t1+t2))}\displaystyle\hskip 100.0pt\sum_{i=1}^{m}{\mathcal{\tau}}_{\hskip 1.0pti}=n-(s+r)\,\bigg{|}\,\sum_{i=1}^{m}{\mathcal{T}}_{\hskip 1.0pti}=t-(t_{1}+t_{2})\big{)}\bigg{\}}
×f𝒯m(t(t1+t2))𝖯{τ0=r,𝒯[r]dt1,𝒳[r]dx1}\displaystyle\hskip 30.0pt\times f_{{\mathcal{T}}_{\hskip 1.0pt}}^{*m}\big{(}t-(t_{1}+t_{2})\big{)}\,\mathsf{P}\Big{\{}{\mathcal{\tau}}_{\hskip 1.0pt0}=r,{\mathcal{T}}_{\hskip 1.0pt}^{[r]}\in dt_{1},{\mathcal{X}}_{\hskip 1.0pt}^{[r]}\in dx_{1}\Big{\}}
×𝖯{τ1s,𝒩(t2)=s,𝒳(s)dx2}dt2.\displaystyle\hskip 30.0pt\times\mathsf{P}\Big{\{}{\mathcal{\tau}}_{\hskip 1.0pt1}\geqslant s,{\mathcal{N}}_{\hskip 1.0pt}(t_{2})=s,{\mathcal{X}}_{\hskip 1.0pt}^{(s)}\in dx_{2}\Big{\}}\,dt_{2}.

It is clear that for large tt only the expression (5.1) matters.

In words, in (5.1) the first incomplete block is made up by rr summands and the value tt cannot be exceeded all over this block because t1<tt_{1}<t; there are mm complete blocks, and exceeding the value tt still cannot happen, but it does occurs on the ssth summand of the the (m+1)(m+1)th complete block consisting of more than ss summands.

Steps  B and C  remain essentially the same as in the proof of Theorem 3.2.

Step  D:  application of refined CLT.  The integrand in (5.1) is ready for application of Edgeworth’s expansion in three–dimensional hybrid integro–local–local CLT with non–uniform remainder term. These theorems see in [31][33].

Steps  E and F  are technically much more complicated than in the proof of refined elementary renewal theorem (3.2) and Theorem 3.2 (they require a number of special identities), but use essentially the same analytical methods. ∎

Remark 5.1.

Theorem 3.2 is a corollary of Theorem 5.1. For /\mathcal{I\hskip-0.5pt/I}  basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, the modular structure is yielded by random indices 00<1<0\equiv\mathbb{\ell}_{0}<\mathbb{\ell}_{1}<\dots and, using Wald’s identities, S\mathcal{M}_{{S}}{\hskip 0.6pt} is transformed into MSM_{{S}}{\,}, and 𝒦(3,0)\mathcal{K}_{(3,0)} is transformed into K(3,0)K_{(3,0)}.∎

5.2. Limit theorems for 𝓡/𝓓\boldsymbol{\mathcal{R\hskip-0.5pt/D}}  sums with Markov dependence

Markov renewal process (see, e.g., [24], [78]) is a homogeneous two–dimensional Markov chain (ξk,Tk)(\xi_{k},T_{k}), k=0,1,k=0,1,\dots, which takes values in a general state space (E×𝖱+,)(\mathit{E}\times\mathsf{R}^{+},\mathcal{E}\otimes\mathcal{R}). Its transition functions are defined by a semi–Markov transition kernel. Markov renewal processes and their counterpart, semi–Markov processes, are among the most popular models of applied probability222222See, e.g., two bibliographies on semi–Markov processes [85] and [86]. The former consists of about 600 papers by some 300 authors and the later of almost a thousand papers by more than 800 authors..

In the same way as for renewal process, this model can be reformulated in terms of /𝒟\mathcal{R\hskip-0.5pt/D}  basis (Ti,Xi)(T_{i},X_{i}), i=1,2,i=1,2,\dots, and /𝒟\mathcal{R\hskip-0.5pt/D}  sum with Markov dependence. Under certain regularity conditions, the sequence 0<𝒥0<𝒥1<0<\mathcal{J}_{0}<\mathcal{J}_{1}<\dots from Definition 5.1 is built in [7], [72]. Theorem analogous to Theorem 5.1 see in [60].

Remark 5.2 (Refined CLT for Markov chains).

In the special case Ti1T_{i}\equiv 1, the /𝒟\mathcal{R\hskip-0.5pt/D}  sum with Markov dependence becomes the ordinary sum Sn{S}_{n} of random variables defined on the Markov chain {ξi}i0\{\xi_{i}\}_{i\geqslant 0}.

Originally, “Dœblin’s trick” was applied to Sn{S}_{n} with discrete {ξi}i0\{\xi_{i}\}_{i\geqslant 0}. Subsequently, it was repeatedly noted (see, e.g., [25] and [68]) that not the full countable space structure is often needed, but just the existence of one single “proper” point. The results then carry over with only notational changes to the countable case and, moreover, to (by means of the “splitting technique” and its counterparts, see [8], [72], [73], and [68]) irreducible recurrent general state space Markov chain {ξi}i0\{\xi_{i}\}_{i\geqslant 0} which satisfies strong minorization condition. Using modular advanced technique, for discrete (see [14]) and general state space Markov chains (see [15], [57], [59]) Berry–Esseen’s estimate, Edgeworth’s expansions, and asymptotic results for large deviations in CLT have been obtained.∎

5.3. Limit theorems for proper 𝓘/𝓓\boldsymbol{\mathcal{I\hskip-0.5pt/D}}  sums with Markov dependence

Markov additive process (see, e.g., [23], [78]) is a generalization of Markov renewal process with (ξk,Tk)(\xi_{k},T_{k}), k=0,1,k=0,1,\dots, taking values in (E×𝖱,)(\mathit{E}\times\mathsf{R},\mathcal{E}\otimes\mathcal{R}) rather than in (E×𝖱+,)(\mathit{E}\times\mathsf{R}^{+},\mathcal{E}\otimes\mathcal{R}). Under certain regularity conditions, for proper /𝒟\mathcal{I\hskip-0.5pt/D}  sums with Markov dependence, the sequence 0<𝒥0<𝒥1<0<\mathcal{J}_{0}<\mathcal{J}_{1}<\dots from Definition 5.1 is built in [4] and [5]. Loosely speaking, these are the ladder points at which regeneration occurs.

6. Analysis of defective 𝓘/𝓓\boldsymbol{\mathcal{I\hskip-0.5pt/D}}  sums

For defective /𝒟\mathcal{I\hskip-0.5pt/D}  sums, a deep analytical insight, such as in the case of defective /\mathcal{I\hskip-0.5pt/I}  sums, is (see, e.g., [70], [74][77]) hardly possible. An alternative is computer–intensive analysis (see, e.g., [2], [50]).

Note that the papers devoted to defective /𝒟\mathcal{I\hskip-0.5pt/D}  sums with Markov dependence are formulated as a study of the ruin probabilities for risk processes of Markovian type. This link with the ruin theory not only suggests rational examples of defective /𝒟\mathcal{I\hskip-0.5pt/D}  sums, but also indicates areas in which they are of major interest.

7. A concluding remark

Due credits must be given to Dœblin (1915–1940), who (see [51], [52], [67]) put forth (see [26][29]) the major innovative ideas of coupling, decomposition and dissection232323The English translation of the title of [26] is “On two problems of Kolmogorov concerning countable Markov chains”, with reference to [48]. Therefore, Dœblin’s glory in this regard is partly shared with Kolmogorov. in Markov chains. The contributions of many scientists after him are mere a development of his seminal ideas. In this regard, compelling is the phrase from [40], that “our goal is to isolate the key ingredients in Dœblin’s proof and use them to derive ergodic theorems for much more general Markov processes on a typically uncountable state space”.

Having created a tool for sequential analysis, Anscombe (see [6]) rediscovered that part of “Dœblin’s trick” (as Kesten called it in [46]), or “Dœblin’s method” (as Kolmogorov called it in [49]), which is the use of Kolmogorov’s inequality for maximum of partial sums. Rényi (see [81], [82]), when he adapted Anscombe’s advance to compound sums SNt{S}_{N_{t}} with NtN_{t} more general242424In Rényi’s sums (see Remark 2.1), NtN_{t} is dependent on i.i.d.  summands and Nt/t𝖯ϱN_{t}/\,t\overset{\scriptscriptstyle{\mathsf{P}}}{\rightarrow}\varrho, tt\to\infty, but is of a general form. than in (2.1), acknowledged (see [82]) that “K.L. Chung kindly called my attention to the fact that the main idea of the proof (the application of the inequality of Kolmogorov \dots) given in [81] is due to Dœblin (see [26]). It has been recently proved \dots that the method in question can lead to the proof of the most general form of Anscombe’s theorem.” Just the way it is in the world of ideas.

References

  • [1]
  • [2] Albrecher, H., and Kantor, J. (2002) Simulation of ruin probabilities for risk processes of Markovian type, Monte Carlo Methods and Appl., 8, 111–127.
  • [3] Alsmeyer, G. (1988) Second–order approximations for certain stopped sums in extended renewal theory, Advances in Applied Probability, 20, 391–410.
  • [4] Alsmeyer, G. (2000) The ladder variables of a Markov random walk, Probability and Mathematical Statistics, 20, 1, 151–168.
  • [5] Alsmeyer, G. (2018) Ladder epochs and ladder chain of a Markov random walk with discrete driving chain, Advances in Applied Probability, 50, 31–46.
  • [6] Anscombe, F.J. (1952) Large–sample theory of sequential estimation, Proc. Cambridge Phil. Soc., 48, 600–607.
  • [7] Athreya, K.B., McDonald, D., and Ney, P. (1978) Limit theorems for semi–Markov processes and renewal theory for Markov chains, Ann. Probab., 6, 788–797.
  • [8] Athreya, K.B., and Ney, P. (1978) A new approach to the limit theory of recurrent Markov chains, Trans. Amer. Math. Soc., 245, 493–501.
  • [9] von Bahr, B. (1974) Ruin probabilities expressed in terms of ladder height distributions, Scandinavian Actuarial Journal, 57, 190–204.
  • [10] Baldwin, C.Y., and Clark, K.B. (2000) Design Rules. Vol. 1: The Power of Modularity. MIT Press, Cambridge, MA.
  • [11] Bhattacharya, R.N., and Ranga Rao, R. (1976) Normal Approximation and Asymptotic Expansions. John Wiley & Sons, New York.
  • [12] Bingham, N.H. (1989) The work of A.N. Kolmogorov on strong limit theorems, Theory Probab. Appl., 34, 1, 129–139.
  • [13] Blackwell, D. (1953) Extension of a renewal theorem, Pacific J. Math., 3, 315–320.
  • [14] Bolthausen, E. (1980) The Berry–Esseen theorem for functionals of discrete Markov chains, Z. Wahrscheinlichkeitstheorie Verw. Geb., B. 54, 59–73.
  • [15] Bolthausen, E. (1982) The Berry–Esseen theorem for strongly mixing Harris recurrent Markov chains, Z. Wahrscheinlichkeitstheorie Verw. Geb., B. 60, 283–289.
  • [16] Brooks, F.P. (1975) The Mythical Man–Month: Essays on Software Engineering. Addison–Wesley, Reading.
  • [17] Brown, M., and Solomon, H. (1975) A second–order approximation for the variance of a renewal reward process, Stochastic Processes and their Applications, 3, 301–314.
  • [18] Chibisov, D.M. (2016) Bernoulli’s law of large numbers and the strong law of large numbers, Theory Probab. Appl., 60, 318–319.
  • [19] Chow, Y.S., and Teicher, H. (1997) Probability Theory. Independence, Interchangeability, Martingales. Springer Texts in Statistics, 3rd ed., Springer, New York.
  • [20] Chung, K.L. (1967) Markov Chains with Stationary Transition Probabilities. Springer, Berlin, Heidelberg, New York.
  • [21] Cox, D.R. (1970) Renewal Theory. Methuen & Co., London.
  • [22] Cox, D.R., and Miller, H.D. (2001) The Theory of Stochastic Processes. Chapman and Hall/CRC, Boca Raton.
  • [23] Çinlar, E. (1972) Markov additive processes. I, II, Probability Theory and Related Fields, 24, 85–93, 95–121.
  • [24] Çinlar, E. (1975) Markov renewal theory: a survey, Management Science, 21, 727–752.
  • [25] Dacunha–Castelle, D., and Duflo, M. (1986) Probability and Statistics. Vol. II. Springer, New York, Berlin, etc.
  • [26] Dœblin, W. (1938) Sur deux problèmes de M. Kolmogoroff concernant les chaînes dénombrables, Bull. Soc. Math. de France, 66, 210–220.
  • [27] Dœblin, W. (1938) Exposé de la théorie des chaînes simples constants de Markoff à un nombre fini d’états, Revue Math. de l’Union Interbalkanique, 2, 77–105.
  • [28] Dœblin, W. (1940) Éléments d’une théorie générale des chaînes simple constantes de Markoff, Ann. Sci. École Norm. Sup., 57, 61–111.
  • [29] Dœblin, W., and Fortet, R. (1937) Sur des chaînes à liaisons complètes, Bull. Soc. Math. de France, 65, 132–148.
  • [30] Doob, J.L. (1948) Renewal theory from the point of view of the theory of probability, Trans. Amer. Math. Soc., 63, 422–438.
  • [31] Dubinskaite, J. (1982) Limit theorems in 𝖱k\mathsf{R}^{k}. I, Lith. Math. J., 22, 129–140.
  • [32] Dubinskaite, J. (1984) Limit theorems in 𝖱k\mathsf{R}^{k}. II, Lith. Math. J., 24, 256–265.
  • [33] Dubinskaite, J. (1984) Limit theorems in 𝖱k\mathsf{R}^{k}. III, Lith. Math. J., 24, 352–334.
  • [34] Feller, W. (1940) On the integro–differential equations of purely discontinuous Markov processes, Trans. Amer. Math. Soc., 48, 488–515.
  • [35] Feller, W. (1941) On the integral equation of renewal theory, Ann. Math. Stat., 12, 243–267.
  • [36] Feller, W. (1949) Fluctuation theory of recurrent events, Trans. Amer. Math. Soc., 67, 98–119.
  • [37] Feller, W. (1971) An Introduction to Probability Theory and its Applications. Vol. II. 2nd ed., John Wiley & Sons, New York, etc.
  • [38] Gnedenko, B.V., and Korolev, V.Yu. (1996) Random Summation. Limit Theorems and Applications. CRC Press: Boca Raton.
  • [39] Gœtze, F., and Hipp, C. (1983) Asymptotic expansions for sums of weakly dependent random variables, Z. Wahrscheinlichkeitstheorie Verw. Geb., 64, 211–239.
  • [40] Griffeath, D. (1978) Coupling methods for Markov processes. Thesis, Cornell Univ. In: Rota, G.–C. (ed.) Studies in Probability and Ergodic Theory. Advances in Mathematics: Supplementary Studies, Vol. 2, 1–43. Academic Press, New York, etc.
  • [41] Gut, A. (2009) Stopped Random Walks. Limit Theorems and Applications. 2nd ed., Springer–Verlag, New York.
  • [42] Hervé, L., and Pène, F. (2010) The Nagaev–Guivarc’h method via the Keller–Liverani theorem, Bull. Soc. Math. de France, 138, 3, 415–489.
  • [43] Hipp, C. (1985) Asymptotic expansions in the central limit theorem for compound and Markov processes, Z. Wahrscheinlichkeitstheorie Verw. Geb., 69, 361–385.
  • [44] Jewell, W.B. (1967) Fluctuations of a renewal–reward process, J. Math. Anal. Appl., 19, 2, 309–329.
  • [45] Karlin, S., and Taylor, H.M. (1975) A First Course in Stochastic Processes. 2nd ed., Academic Press, New York, etc.
  • [46] Kesten, H. (1977) Book review (of [80]), Bull. Amer. Math. Soc., 83, 696–697.
  • [47] Kingman, J.F.C. (1972) Regenerative phenomena. John Wiley & Sons, London, etc.
  • [48] Kolmogorov, A.N. (1936) Anfangsgründe der Theorie der Markoffschen Ketten mit unendlich vielen möglichen Zuständen, Mat. Sbornik, N. Ser., 1, 607–610.
  • [49] Kolmogorov, A.N. (1949) A local limit theorem for classical Markov chains, Izv. Akad. Nauk SSSR, Ser. Mat. 13, 281–300 (in Russian).
  • [50] Lehtonen, T., and Nyrhinen, H. (1992) On asymptotically efficient simulation of ruin probabilities in a Markovian environment, Scandinavian Actuarial Journal, 60–75.
  • [51] Lévy, P. (1955) W. Dœblin (V. Doblin) (1915–1940), Rev. Histoire Sci. Appl., 8, 107–115.
  • [52] Lindvall, T. (1991) W. Dœblin, 1915–1940, Ann. Probab., 19, 929–934.
  • [53] Loève, M. (1950) Fundamental limit theorems of probability theory, Ann. Math. Stat., 21, 321–338.
  • [54] Lotka, A. (1939) A contribution to the theory of self–renewing aggregates, with special reference to industrial replacement, Ann. Math. Stat., 10, 1–25.
  • [55] Malinovskii, V.K. (1984) On asymptotic expansions in the central limit theorem for Harris recurrent Markov chains, Soviet Math. Dokl., 29, 679–684.
  • [56] Malinovskii, V.K. (1985) On some asymptotic relations and identities for Harris recurrent Markov chains. In: Statistics and Control of Stochastic Processes, Optimization Software, 317–336.
  • [57] Malinovskii, V.K. (1987) Limit theorems for Harris Markov chains. I, Theory Probab. Appl., 31, 269–285.
  • [58] Malinovskii, V.K. (1988) On a limit theorem for dependent random variables, Ann. Academiæ Scientiarum Fennicæ, 13, 225–229.
  • [59] Malinovskii, V.K. (1990) Limit theorems for Harris Markov chains. II, Theory Probab. Appl., 34, 252–265.
  • [60] Malinovskii, V.K. (1991) On integral and local limit theorems for recursive252525Erroneous translation into English. In the original paper published in 1988 in Russian: “recurrent”. Markov renewal processes, Journal of Soviet Math., 57, 4, 3286–3301. [MR 91m: 60161] Translated from: Problems of Stability of Stoch. Models (1988), Ed. V.M. Zolotarev, VNIISI, Moscow, 100–115 (in Russian).
  • [61] Malinovskii, V.K. (1992) Asymptotic expansions in the central limit theorem for stopped random walks, Theory Probab. Appl., 36, 827–829.
  • [62] Malinovskii, V.K. (1994) Limit theorems for stopped random sequences I. Rates of convergence and asymptotic expansions, Theory Probab. Appl., 38, 673–693.
  • [63] Malinovskii, V.K. (1994) Corrected normal approximation for the probability of ruin within finite time, Scandinavian Actuarial Journal, 161–174.
  • [64] Malinovskii, V.K. (2021) Insurance Planning Models. World Scientific, Singapore.
  • [65] Malinovskii, V.K. (2021) Level–Crossing Problems and Inverse Gaussian Distributions. Chapman and Hall/CRC, Boca Raton.
  • [66] Malinovskii, V.K. (2021) Risk Measures and Insurance Solvency Benchmarks. Chapman and Hall/CRC, Boca Raton.
  • [67] Meyn, S.P., and Tweedie, R.L. (1993) The Dœblin decomposition, Contemporary Mathematics, 149, 211–225.
  • [68] Meyn, S.P., and Tweedie, R.L. (2009) Markov Chains and Stochastic Stability. 2nd ed., Cambridge Univ. Press, Cambridge, etc.
  • [69] Mikes, G. (1970) How to Be an Alien: A Handbook for Beginners and Advanced Pupils. Gardners Books.
  • [70] Mikosch, T., and Samorodnitsky, G. (2000) Ruin probability with claims modeled by a stationary ergodic stable process, Ann. Probab., 28, 1814–1851.
  • [71] Nummelin, E. (1978) A splitting technique for Harris recurrent chains, Z. Wahrscheinlichkeitstheorie Verw. Geb., 43, 309–318.
  • [72] Nummelin, E. (1978) Uniform and ratio limit theorems for Markov renewal and semi–regenerative processes on a general state space, Ann. Inst. H. Poincare, Sect. B, Vol. XIV, 119–143.
  • [73] Nummelin, E. (1984) General Irreducible Markov Chains and Non–negative Operators. Cambridge Univ. Press, Cambridge, etc.
  • [74] Nyrhinen, H. (1998) Rough descriptions of ruin for a general class of surplus processes, Advances in Applied Probability, 30, 1008–1026.
  • [75] Nyrhinen, H. (1999) Large deviations for the time of ruin, Journal of Applied Probability, 36, 733–746.
  • [76] Nyrhinen, H. (1999) On the ruin probabilities in a general economic environment, Stochastic Processes and their Applications, 83, 319–330.
  • [77] Nyrhinen, H. (2001) Finite and infinite time ruin probabilities in a stochastic economic environment, Stochastic Processes and their Applications, 92, 265–285.
  • [78] Pacheco, A., Prabhu, N.U., and Tang, L.C. (2009) Markov–modulated Processes and Semiregenerative Phenomena. World Scientific, New Jersey, etc.
  • [79] Parnas, D.L. (1972) On the criteria to be used in decomposing systems into modules, Communications of the Association for Computing Machinery (ACM), 15, 1053–1058.
  • [80] Petrov, V.V. (1975) Sums of Independent Random Variables. Springer, Berlin, etc.
  • [81] Rényi, A. (1957) On the asymptotic distribution of the sum of a random number of independent random variables, Acta Math. Acad. Sci. Hungar., 8, 193–199.
  • [82] Rényi, A. (1960) On the central limit theorem for the sum of a random number of independent random variables, Acta Math. Acad. Sci. Hungar., 11, 97–102.
  • [83] Rio, E. (2017) Asymptotic Theory of Weakly Dependent Random Processes. Springer, New York.
  • [84] Robbins, H. (1948) The asymptotic distribution of the sum of a random number of random variables, Bull. Amer. Math. Soc., 54, 1151–1161.
  • [85] Teugels, J.L. (1976) A bibliography on semi–Markov processes, Journal of Computational and Applied Mathematics, 2, 125–144.
  • [86] Teugels, J.L. (1986) A second bibliography on semi–Markov processes, In: Semi–Markov Models. Theory and Applications, ed. J. Janssen, 507–584, Plenum Press, New York.
  • [87] Tikhomirov, A.N. (1981) On the convergence rate in the central limit theorem for weakly dependent random variables, Theory Probab. Appl., 25, 790–809.
  • [88] Wolff, R.W. (1989) Stochastic Modeling and the Theory of Queues. Prentice–Hall.