This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Modified Galton-Watson processes with immigration under an alternative offspring mechanism

Wagner Barreto-Souza111Email: wagner.barretosouza@kaust.edu.sa (Corresponding Author),   Sokol Ndreca222Email: sokol@est.ufmg.br,   Rodrigo B. Silva333Email: rodrigo@de.ufpb.br   and Roger W.C. Silva444Email: rogerwcs@est.ufmg.br

Statistics Program, King Abdullah University of Science and Technology, Thuwal, Saudi Arabia
Departamento de Estatística, Universidade Federal de Minas Gerais, Belo Horizonte, Brazil
Departamento de Estatística, Universidade Federal da Paraíba, João Pessoa, Brazil
Abstract

We propose a novel class of count time series models alternative to the classic Galton-Watson process with immigration (GWI) and Bernoulli offspring. A new offspring mechanism is developed and its properties are explored. This novel mechanism, called geometric thinning operator, is used to define a class of modified GWI (MGWI) processes, which induces a certain non-linearity to the models. We show that this non-linearity can produce better results in terms of prediction when compared to the linear case commonly considered in the literature. We explore both stationary and non-stationary versions of our MGWI processes. Inference on the model parameters is addressed and the finite-sample behavior of the estimators investigated through Monte Carlo simulations. Two real data sets are analyzed to illustrate the stationary and non-stationary cases and the gain of the non-linearity induced for our method over the existing linear methods. A generalization of the geometric thinning operator and an associated MGWI process are also proposed and motivated for dealing with zero-inflated or zero-deflated count time series data.

Keywords: Autocorrelation; Count time series; Estimation; INAR processes; Galton-Watson processes; Geometric thinning operator.

1 Introduction

The Galton-Watson (GW) or branching process is a simple and well-used model for describing populations evolving in time. It is defined by a sequence of non-negative integer-valued random variables {Xt}t\{X_{t}\}_{t\in\mathbb{N}} satisfying

Xt=k=1Xt1ζt,k,t,\displaystyle X_{t}=\sum_{k=1}^{X_{t-1}}\zeta_{t,k},\quad t\in\mathbb{N}, (1)

with X0=1X_{0}=1 by convention, where {ζt,k}t,k\{\zeta_{t,k}\}_{t,k\in\mathbb{N}} is a doubly infinite array of independent and identically distributed (iid) random variables. Write FF for the offspring distribution, so that ζt,kF\zeta_{t,k}\sim F for all t,k1t,k\geq 1. In the populational context, the random variable XtX_{t} denotes the size of the tt-th generation. A generalization of this model is obtained by allowing an independent immigration process in (1), which is known as the GW process with immigration (GWI) and given by

Xt=k=1Xt1ζt,k+ϵt,t,\displaystyle X_{t}=\sum_{k=1}^{X_{t-1}}\zeta_{t,k}+\epsilon_{t},\quad t\in\mathbb{N}, (2)

where {ϵt}t\{\epsilon_{t}\}_{t\in\mathbb{N}} is assumed to be a sequence of iid non-negative integer-valued random variables, with ϵt\epsilon_{t} independent of Xs1X_{s-1} and ζs,k\zeta_{s,k}, for all k1k\geq 1 and for all sts\leq t. If one assumes αE(ζs,k)<\alpha\equiv E(\zeta_{s,k})<\infty and μϵE(ϵt)<\mu_{\epsilon}\equiv E(\epsilon_{t})<\infty, then the conditional expectation of the size of the tt-th generation given the size of the (t1)(t-1)-th generation, is linear on Xt1X_{t-1} and given by

E(Xt|Xt1)=αXt1+μϵ.\displaystyle E(X_{t}|X_{t-1})=\alpha X_{t-1}+\mu_{\epsilon}. (3)

An interesting example appears when the offspring is Bernoulli distributed, that is, when P(ζ=1)=1P(ζ=0)=α(0,1)\operatorname{P}\,(\zeta=1)=1-\operatorname{P}\,(\zeta=0)=\alpha\in(0,1). This yields the binomial thinning operator “\circ” by Steutel and van Harn (1979), which is defined by

αXt1k=1Xt1ζt,k.\alpha\circ X_{t-1}\equiv\sum_{k=1}^{X_{t-1}}\zeta_{t,k}.

In this case, the GWI process in (2) is related to the first-order Integer-valued AutoRegressive (INAR) models presented in Alzaid and Al-Osh (1987), McKenzie (1988), and Dion et al. (1995). Conditional least squares estimation for the GWI/INAR models were explored, for instance, by Wei and Winnicki (1990), Ispány et al. (2003), Freeland and McCabe (2005), and Rahimov (2008).

Alternative integer-valued processes based on non-additive innovation through maximum and minimum operations were proposed by Littlejohn (1992), Littlejohn (1996), Kalamkar (1995), Scotto et al. (2016), and Aleksić and Ristić (2021). For the count processes {Xt}t\{X_{t}\}_{t\in\mathbb{N}} considered in these works, a certain non-linearity is induced in the sense that the conditional expectation E(Xt|Xt1)E(X_{t}|X_{t-1}) is non-linear on Xt1X_{t-1} (and also the conditional variance) in contrast with (3). We refer to these models as “non-linear” along with this paper. On the other hand, the immigration interpretation in a populational context is lost due to the non-additive innovation assumption.

Our aim in this paper is to introduce an alternative model to the classic GW process with immigration (GWI) and Bernoulli offspring. We develop a modified GWI process (MGWI) based on a new thinning operator/offspring mechanism while preserving the additive innovation, which has a practical interpretation. We show that this new mechanism, called the geometric thinning operator, induces a certain non-linearity when compared to the classic GWI/INAR processes. We now highlight other contributions of the present paper:

  • (i)

    development of inferential procedures and numerical experiments, which are not well-explored for the existing non-linear models aforementioned;

  • (ii)

    properties of the novel geometric thinning operator are established;

  • (iii)

    a particular MGWI process with geometric marginals is investigated in detail, including an explicit expression for the autocorrelation function;

  • (iv)

    both stationary and non-stationary cases are explored, being the last important for allowing the inclusion of covariates, a feature not considered by the current non-linear models;

  • (v)

    empirical evidences that the non-linearity induced for our MGWI processes can produce better results in terms of prediction when compared to the linear case (commonly considered in the literature);

  • (vi)

    a generalization of the geometric thinning operator and an associated MGWI process are also proposed and motivated for dealing with zero-inflated or zero-deflated count time series data.

The paper is organized as follows. In Section 2, we introduce the new geometric thinning operator and explore its properties. Section 3 is devoted to the development of the modified Galton-Watson processes with immigration based on the new operator, with a focus on the case where the marginals are geometrically distributed. Two methods for estimating the model parameters are discussed in Section 4, including Monte Carlo simulations to evaluate the proposed estimators. In Section 5, we introduce a non-stationary MGWI process allowing for the inclusion of covariates and provide some Monte Carlo studies. Section 6 is devoted to two real data applications. Finally, in Section 7, we develop a generalization of the geometric thinning operator and an associated modified GWI model.

2 Geometric thinning operator: definition and properties

In this section, we introduce a new thinning operator and derive its main properties. We begin by introducing some notation. For two random variables XX and YY, we write min{X,Y}:=XY\min\{X,Y\}:=X\wedge Y to denote the minimum between XX and YY. The probability generating function (pgf) of a non-negative integer-valued random variable YY is denoted by

ΨY(s)=E(sY)=k=0skP(Y=k),\displaystyle\Psi_{Y}(s)=\operatorname{E}\left(s^{Y}\right)=\sum_{k=0}^{\infty}s^{k}\operatorname{P}\,(Y=k),

for all values of ss for which the right-hand side converges absolutely. The nn-th derivative of ΨY(x)\Psi_{Y}(x) with respect to xx and evaluated at x=x0x=x_{0} is denoted by ΨY(n)(x0)\Psi_{Y}^{(n)}(x_{0}).

Let ZZ be a geometric random variable with parameter α>0\alpha>0 and probability function assuming the form

P(Z=k)=αk(1+α)k+1,k=0,1,.\operatorname{P}\,(Z=k)=\frac{\alpha^{k}}{(1+\alpha)^{k+1}},\quad k=0,1,\dots.

In this case, the pgf of XX is

ΨZ(s)=11+α(1s),|s|<1+α1,\displaystyle\Psi_{Z}(s)=\frac{1}{1+\alpha(1-s)},\quad|s|<1+\alpha^{-1}, (4)

and the parameter α\alpha has the interpretation α=E(Z)>0\alpha=E(Z)>0. The shorthand notation ZGeo(α)Z\sim\mathrm{Geo}(\alpha) will be used throughout the text. We are ready to introduce the new operator and explore some of its properties.

Definition 1.

(Geometric thinning operator) Let XX be a non-negative integer-valued random variable, independent of ZGeo(α)Z\sim\mathrm{Geo}(\alpha), with α>0\alpha>0. The geometric thinning operator \triangle is defined by

αXmin(X,Z).\alpha\operatorname{\triangle}X\equiv\min\left(X,Z\right). (5)
Remark 1.

The operator \operatorname{\triangle} defined in (5) satisfies αXX\alpha\operatorname{\triangle}X\leq X, like the classic binomial thinning operator \circ. Therefore, \operatorname{\triangle} is indeed a thinning operator.

In what follows, we present some properties of the proposed geometric thinning operator. We start by obtaining its probability generating function.

Proposition 1.

Let XX be a non-negative integer-valued random variable with pgf ΨX\Psi_{X}. Then, the pgf of αX\alpha\operatorname{\triangle}X is given by

ΨαX(s)=1+α(1s)ΨX(αs1+α)1+α(1s),|s|<1+α1.\displaystyle\Psi_{\alpha\operatorname{\triangle}X}(s)=\frac{1+\alpha(1-s)\Psi_{X}\left(\dfrac{\alpha s}{1+\alpha}\right)}{1+\alpha(1-s)},\quad|s|<1+\alpha^{-1}.
Proof.

By the independence assumption between XX and ZZ, it holds that

P(αX=k)\displaystyle\operatorname{P}\,(\alpha\operatorname{\triangle}X=k) =P(αXk)P(αXk+1)\displaystyle=\operatorname{P}\,(\alpha\operatorname{\triangle}X\geq k)-\operatorname{P}\,(\alpha\operatorname{\triangle}X\geq k+1)
=P(Zk)P(Xk)P(Zk+1)P(Xk+1)\displaystyle=\operatorname{P}\,(Z\geq k)\operatorname{P}\,(X\geq k)-\operatorname{P}\,(Z\geq k+1)\operatorname{P}\,(X\geq k+1)
=(α1+α)k[P(X=k)+11+αP(Xk+1)].\displaystyle=\left(\frac{\alpha}{1+\alpha}\right)^{k}\left[\operatorname{P}\,(X=k)+\frac{1}{1+\alpha}\operatorname{P}\,(X\geq k+1)\right].

Hence,

ΨαX(s)\displaystyle\Psi_{\alpha\operatorname{\triangle}X}(s) =k=0(αs1+α)kP(X=k)+11+αk=0(αs1+α)kP(Xk+1)\displaystyle=\sum_{k=0}^{\infty}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\operatorname{P}\,(X=k)+\frac{1}{1+\alpha}\sum_{k=0}^{\infty}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\operatorname{P}\,(X\geq k+1)
=ΨX(αs1+α)11+αΨX(αs1+α)+11+αk=0(αs1+α)kP(Xk)\displaystyle=\Psi_{X}\left(\frac{\alpha s}{1+\alpha}\right)-\frac{1}{1+\alpha}\Psi_{X}\left(\frac{\alpha s}{1+\alpha}\right)+\frac{1}{1+\alpha}\sum_{k=0}^{\infty}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\operatorname{P}\,(X\geq k)
=α1+αΨX(αs1+α)+11+αk=0(αs1+α)kP(Xk).\displaystyle=\frac{\alpha}{1+\alpha}\Psi_{X}\left(\frac{\alpha s}{1+\alpha}\right)+\frac{1}{1+\alpha}\sum_{k=0}^{\infty}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\operatorname{P}\,(X\geq k).

The second term on the last equality can be expressed as

11+αk=0(αs1+α)kP(Xk)\displaystyle\frac{1}{1+\alpha}\sum_{k=0}^{\infty}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\operatorname{P}\,(X\geq k) =11+αk=0(αs1+α)kl=kP(X=l)\displaystyle=\frac{1}{1+\alpha}\sum_{k=0}^{\infty}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\sum_{l=k}^{\infty}\operatorname{P}\,(X=l)
=11+αl=0k=0l(αs1+α)kP(X=l)\displaystyle=\frac{1}{1+\alpha}\sum_{l=0}^{\infty}\sum_{k=0}^{l}\left(\frac{\alpha s}{1+\alpha}\right)^{k}\operatorname{P}\,(X=l)
=11+ααs[1αs1+αΨX(αs1+α)].\displaystyle=\frac{1}{1+\alpha-\alpha s}\left[1-\frac{\alpha s}{1+\alpha}\Psi_{X}\left(\frac{\alpha s}{1+\alpha}\right)\right].

The result follows by rearranging the terms. ∎

The next result gives us the moments of αX\alpha\operatorname{\triangle}X, which will be important to discuss prediction and forecasting in what follows.

Proposition 2.

Let \triangle be the geometric thinning operator in (5). It holds that the nn-th factorial moment of αX\alpha\operatorname{\triangle}X is given by

E((αX)n)=n!αn{1k=0n1ΨX(k)(α1+α)k!(1+α)k},\operatorname{E}\,\big{(}(\alpha\operatorname{\triangle}X)_{n}\big{)}=n!\alpha^{n}\left\{1-\sum_{k=0}^{n-1}\dfrac{\Psi_{X}^{(k)}\left(\frac{\alpha}{1+\alpha}\right)}{k!(1+\alpha)^{k}}\right\},

for nn\in\mathbb{N}, where (αX)nαX×(αX1)××(αXn+1)(\alpha\operatorname{\triangle}X)_{n}\equiv\alpha\operatorname{\triangle}X\times(\alpha\operatorname{\triangle}X-1)\times\ldots\times(\alpha\operatorname{\triangle}X-n+1).

Proof.

The result follows by using the pgf given in Proposition 1 and the generalized Leibniz rule for derivatives, namely (d1d2)(n)(s)=k=0n(nk)d1(nk)(s)d2(k)(s)(d_{1}d_{2})^{(n)}(s)=\sum_{k=0}^{n}\binom{n}{k}d_{1}^{(n-k)}(s)d_{2}^{(k)}(s), with d1(s)=1+α(1s)ΨX(αs1+α)d_{1}(s)=1+\alpha(1-s)\Psi_{X}\left(\dfrac{\alpha s}{1+\alpha}\right) and d2(s)=11+α(1s)d_{2}(s)=\dfrac{1}{1+\alpha(1-s)}. ∎

In what follows, the notation XYX\Rightarrow Y means XX weakly converges to YY.

Proposition 3.

Let \triangle be the geometric thinning operator in (5). Then,

  • (i)

    αX0,as α0,\alpha\operatorname{\triangle}X\Rightarrow 0,\quad\text{as }\alpha\to 0,

  • (ii)

    αXX,as α.\alpha\operatorname{\triangle}X\Rightarrow X,\quad\text{as }\alpha\to\infty.

Proof.

The proof follows immediately from Proposition 1 and the Continuity Theorem for pgf’s. ∎

We now show a property of the operator \operatorname{\triangle} of own interest.

Proposition 4.

Let Z1,,ZnZ_{1},\ldots,Z_{n} be independent geometric random variables with parameters α1,,αn\alpha_{1},\ldots,\alpha_{n}, respectively. Assume that X1,,XnX_{1},\ldots,X_{n} are non-negative integer-valued random variables independent of the ZZ’s, and let αiXi=min(Xi,Zi)\alpha_{i}\operatorname{\triangle}X_{i}=\min\left(X_{i},Z_{i}\right). Then,

k=1nαkXk=α~nk=1nXk,\displaystyle\wedge_{k=1}^{n}\alpha_{k}\operatorname{\triangle}X_{k}=\widetilde{\alpha}_{n}\operatorname{\triangle}\wedge_{k=1}^{n}X_{k}, (6)

with α~n=k=1nαkk=1n(1+αk)k=1nαk\widetilde{\alpha}_{n}=\dfrac{\prod_{k=1}^{n}\alpha_{k}}{\prod_{k=1}^{n}(1+\alpha_{k})-\prod_{k=1}^{n}\alpha_{k}}, nn\in\mathbb{N}.

Proof.

We prove (6) by induction on nn. For n=2n=2, it holds that

k=12αkXk=k=12(XkZk)=(X1X2)(Z1Z2)=α~2k=12Xk,\wedge_{k=1}^{2}\alpha_{k}\operatorname{\triangle}X_{k}=\wedge_{k=1}^{2}(X_{k}\wedge Z_{k})=(X_{1}\wedge X_{2})\wedge(Z_{1}\wedge Z_{2})=\widetilde{\alpha}_{2}\operatorname{\triangle}\wedge_{k=1}^{2}X_{k},

where α~2=k=12αkk=12(1+αk)k=12αk\widetilde{\alpha}_{2}=\dfrac{\prod_{k=1}^{2}\alpha_{k}}{\prod_{k=1}^{2}(1+\alpha_{k})-\prod_{k=1}^{2}\alpha_{k}}. Assume that k=1n1αkXk=α~n1k=1n1Xk\wedge_{k=1}^{n-1}\alpha_{k}\operatorname{\triangle}X_{k}=\widetilde{\alpha}_{n-1}\operatorname{\triangle}\wedge_{k=1}^{n-1}X_{k}. Since

k=1nαkXk=(k=1n1αkXk)(αnXn)=(α~n1k=1n1Xk)(αnXn)=α~nk=1nXk,\wedge_{k=1}^{n}\alpha_{k}\operatorname{\triangle}X_{k}=(\wedge_{k=1}^{n-1}\alpha_{k}\operatorname{\triangle}X_{k})\wedge(\alpha_{n}\operatorname{\triangle}X_{n})=(\,\widetilde{\alpha}_{n-1}\operatorname{\triangle}\wedge_{k=1}^{n-1}X_{k})\wedge(\alpha_{n}\operatorname{\triangle}X_{n})=\widetilde{\alpha}_{n}\operatorname{\triangle}\wedge_{k=1}^{n}X_{k},

the proof is complete. ∎

In the next section, we introduce our class of modified GW processes and provide some of their properties.

3 Modified Galton-Watson processes with immigration

In this section, we introduce modified GW processes with immigration based on the new geometric thinning operator \operatorname{\triangle} defined in Section 2 and explore a special case when the marginals are geometrically distributed.

Definition 2.

A sequence of random variables {Xt}t\{X_{t}\}_{t\in\mathbb{N}} is said to be a modified GW process with immigration (in short MGWI) if it satisfies the stochastic equation

Xt=αXt1+ϵt,t,X_{t}=\alpha\operatorname{\triangle}X_{t-1}+\epsilon_{t},\quad t\in\mathbb{N}, (7)

with αXt1=min(Xt1,Zt)\alpha\operatorname{\triangle}X_{t-1}=\min\left(X_{t-1},Z_{t}\right), {Zt}t\left\{Z_{t}\right\}_{t\in\mathbb{N}} being a sequence of iid random variables with Z1Geo(α)Z_{1}{\sim}\mbox{Geo}(\alpha), {ϵt}t\{\epsilon_{t}\}_{t\in\mathbb{N}} being an iid non-negative integer-valued random variables called innovations, where ϵt\epsilon_{t} is independent of XtlX_{t-l} and Ztl+1Z_{t-l+1}, for all l1l\geq 1, and X0X_{0} is some starting value/random variable.

The following theorem is an important result concerning the modified GWI process.

Theorem 5.

The stochastic process {Xt}t\{X_{t}\}_{t\in\mathbb{N}} in (7) is stationary and ergodic.

Proof.

Consider the process {Wt}t={(Zt,ϵt)}t\{W_{t}\}_{t\in\mathbb{N}}=\{(Z_{t},\epsilon_{t})\}_{t\in\mathbb{N}}. Since this is a sequence of iid bivariate vectors, it follows that {Wt}t1\{W_{t}\}_{t\geq 1} is stationary and ergodic. Now, note that there is a real function ξ\xi, which does not depend on tt, such that

Xt=ξ((Zs,ϵs), 1st).X_{t}=\xi\left((Z_{s},\epsilon_{s}),\,1\leq s\leq t\right).

Hence, the result follows by applying Theorem 36.4 from Billingsley (1995). ∎

From now on, we focus our attention on a special case from our class of MGWI processes when the marginals are geometrically distributed. To do this, let us first discuss the zero-modified geometric (ZMG) distribution, which will play an important role in our model construction.

We say that a random variable YY follows a ZMG distribution with parameters μ>0\mu>0 and π(1/μ,1)\pi\in(-1/\mu,1) if its probability function is given by

P(Y=k)={π+(1π)1(1+μ),for k=0,(1π)μk(1+μ)k+1,for k=1,2,.\displaystyle\operatorname{P}\,(Y=k)=\left\{\begin{array}[]{ll}\pi+(1-\pi)\dfrac{1}{(1+\mu)},&\text{for }k=0,\\ (1-\pi)\dfrac{\mu^{k}}{(1+\mu)^{k+1}},&\text{for }k=1,2,\ldots.\end{array}\right.

We denote YZMG(π,μ)Y\sim\mathrm{ZMG}(\pi,\mu). The geometric distribution with mean μ\mu is obtained as a particular case when π=0\pi=0. For π<0\pi<0 and π>0\pi>0, the ZMG distribution is zero-deflated or zero-inflated with relation to the geometric distribution, respectively. The associated pgf assumes the form

ΨY(s)=1+πμ(1s)1+μ(1s),|s|<1+μ1.\displaystyle\Psi_{Y}(s)=\frac{1+\pi\mu(1-s)}{1+\mu(1-s)},\quad|s|<1+\mu^{-1}. (9)

Now, assume that XGeo(μ)X\sim\mathrm{Geo}(\mu), with μ>0\mu>0. We have that

P(αX>z)\displaystyle\operatorname{P}\,(\alpha\operatorname{\triangle}X>z) =P(X>z)P(Z>z)=[(μ1+μ)(α1+α)]z+1,z=0,1,,\displaystyle=\operatorname{P}\,(X>z)\operatorname{P}\,(Z>z)=\left[\left(\frac{\mu}{1+\mu}\right)\left(\frac{\alpha}{1+\alpha}\right)\right]^{z+1},\quad z=0,1,\dots,

which means αXGeo(αμ1+α+μ)\alpha\operatorname{\triangle}X\sim\mathrm{Geo}\left(\dfrac{\alpha\mu}{1+\alpha+\mu}\right). From (7), it follows that a MGWI process with geometric marginals is well-defined if the function Ψϵ1(s)ΨX(s)ΨαX(s)\Psi_{\epsilon_{1}}(s)\equiv\dfrac{\Psi_{X}(s)}{\Psi_{\alpha\operatorname{\triangle}X}(s)} is a proper pgf, with ss belonging to some interval containing the value 1, where ΨX(s)\Psi_{X}(s) and ΨαX(s)\Psi_{\alpha\operatorname{\triangle}X}(s) are the pgf’s of geometric distributions with means μ\mu and αμ1+α+μ\dfrac{\alpha\mu}{1+\alpha+\mu}, respectively. More specifically, we have

Ψϵ1(s)=1+α1+μ+αμ(1s)1+μ(1s),|s|<1+μ1,\Psi_{\epsilon_{1}}(s)=\dfrac{1+\frac{\alpha}{1+\mu+\alpha}\mu(1-s)}{1+\mu(1-s)},\quad|s|<1+\mu^{-1}, (10)

which corresponds to the pgf of a zero-modified geometric distribution with parameters μ\mu and α/(1+μ+α)\alpha/(1+\mu+\alpha); see (9). This enables us to define a new geometric process as follows.

Definition 3.

The stationary geometric MGWI (Geo-MGWI) process {Xt}t\{X_{t}\}_{t\in\mathbb{N}} is defined by assuming that (7) holds with {ϵt}tiidZMG(α1+μ+α,μ)\{\epsilon_{t}\}_{t\in\mathbb{N}}\stackrel{{\scriptstyle iid}}{{\sim}}\mathrm{ZMG}\left(\dfrac{\alpha}{1+\mu+\alpha},\mu\right) and X0Geo(μ)X_{0}\sim\mathrm{Geo}(\mu).

From (10), we have that the mean and variance of the innovations {ϵt}t1\{\epsilon_{t}\}_{t\geq 1} are given by

μϵ:=E(ϵt)=μ(1+μ)1+μ+αandσϵ2:=Var(ϵt)=μ(1+μ)1+μ+α[1+μ(1+μ+2α)1+μ+α],\mu_{\epsilon}:=\operatorname{E}\,(\epsilon_{t})=\frac{\mu(1+\mu)}{1+\mu+\alpha}\quad\text{and}\quad\sigma_{\epsilon}^{2}:=\operatorname{Var}\,(\epsilon_{t})=\frac{\mu(1+\mu)}{1+\mu+\alpha}\left[1+\frac{\mu(1+\mu+2\alpha)}{1+\mu+\alpha}\right],

respectively. Additionally, the third and forth moments of the innovations are

E(ϵt3)=μ(1+μ)1+μ+α(6μ2+4μ+1)andE(ϵt4)=μ(1+μ)1+μ+α(24μ3+36μ2+12μ+5).\displaystyle E(\epsilon_{t}^{3})=\dfrac{\mu(1+\mu)}{1+\mu+\alpha}(6\mu^{2}+4\mu+1)\quad\mbox{and}\quad E(\epsilon_{t}^{4})=\dfrac{\mu(1+\mu)}{1+\mu+\alpha}(24\mu^{3}+36\mu^{2}+12\mu+5).

In what follows, we assume that {Xt}t\{X_{t}\}_{t\in\mathbb{N}} is a Geo-MGWI process and explore some of its properties. We start with the 1-step transition probabilities.

Proposition 6.

The 1-step transition probabilities of the MGWI process, say P(x,y)P(Xt=y|Xt1=x)\operatorname{P}\,(x,y)\equiv\operatorname{P}\,(X_{t}=y\,|\,X_{t-1}=x), assumes the form

P(x,y)={k=0x1P(Z=k)P(ϵt=yk)+P(Zx)P(ϵt=yx),forxy,k=0yP(Z=k)P(ϵt=yk),forx>y,\operatorname{P}\,(x,y)=\begin{cases}\displaystyle\sum_{k=0}^{x-1}\operatorname{P}\,(Z=k)\operatorname{P}\,(\epsilon_{t}=y-k)+\operatorname{P}\,(Z\geq x)\operatorname{P}\,(\epsilon_{t}=y-x),&\text{for}\ \,x\leq y,\\ \displaystyle\sum_{k=0}^{y}\operatorname{P}\,(Z=k)\operatorname{P}\,(\epsilon_{t}=y-k),&\text{for}\ \,x>y,\\ \end{cases} (11)

for x,y=0,1,x,y=0,1,\dots. In particular, we have P(0,y)=P(ϵt=y)\operatorname{P}\,(0,y)=\operatorname{P}\,(\epsilon_{t}=y).

Proof.

For x=0x=0, we have that P(0,y)=P(α0+ϵt=y)=P(ϵt=y)\operatorname{P}\,(0,y)=\operatorname{P}\,(\alpha\operatorname{\triangle}0+\epsilon_{t}=y)=\operatorname{P}\,(\epsilon_{t}=y). For x>0x>0, it follows that

P(x,y)=P(αx+ϵ=y)=k=0yP(αx=k)P(ϵ=yk),\operatorname{P}\,(x,y)=\operatorname{P}\,(\alpha\operatorname{\triangle}x+\epsilon=y)=\sum_{k=0}^{y}\operatorname{P}\,(\alpha\operatorname{\triangle}x=k)\operatorname{P}\,(\epsilon=y-k),

where

P(αx=z)={0,forx<z,P(Zz),forx=z,P(Z=z),forx>z.\operatorname{P}\,(\alpha\operatorname{\triangle}x=z)=\begin{cases}0,&\text{for}\ \,x<z,\\ \operatorname{P}\,(Z\geq z),&\text{for}\ \,x=z,\\ \operatorname{P}\,(Z=z),&\text{for}\ \,x>z.\\ \end{cases}

This gives the desired transition probabilities in (11). ∎

Proposition 7.

The joint pgf of the discrete random vector (Xt,Xt1)(X_{t},X_{t-1}) is given by

ΨXt,Xt1(s1,s2)=Ψϵ(s1)1α(s11)[ΨX(s2)α(s11)ΨX(s1s2α1+α)],\Psi_{X_{t},X_{t-1}}(s_{1},s_{2})=\frac{\Psi_{\epsilon}(s_{1})}{1-\alpha(s_{1}-1)}\left[\Psi_{X}(s_{2})-\alpha(s_{1}-1)\Psi_{X}\left(\frac{s_{1}s_{2}\alpha}{1+\alpha}\right)\right], (12)

with ΨX()\Psi_{X}(\cdot) and Ψϵ()\Psi_{\epsilon}(\cdot) as in (4) and (10), respectively, where s1s_{1} and s2s_{2} belong to some intervals containing the value 1.

Proof.

We have that

ΨXt,Xt1(s1,s2)=E(s1Xts2Xt1)=E(s1αXt1+ϵts2Xt1)=Ψϵt(s1)E(s2Xt1E(s1αXt1|Xt1)),\Psi_{X_{t},X_{t-1}}(s_{1},s_{2})=\operatorname{E}\,\left(s_{1}^{X_{t}}s_{2}^{X_{t-1}}\right)=\operatorname{E}\,\left(s_{1}^{\alpha\operatorname{\triangle}X_{t-1}+\epsilon_{t}}s_{2}^{X_{t-1}}\right)=\Psi_{\epsilon_{t}}(s_{1})\operatorname{E}\,\left(s_{2}^{X_{t-1}}\operatorname{E}\,\left(s_{1}^{\alpha\operatorname{\triangle}X_{t-1}}\,|\,X_{t-1}\right)\right),

where

E(s1αX|X=x)\displaystyle\operatorname{E}\left(s_{1}^{\alpha\operatorname{\triangle}X}\,|\,X=x\right) =k=0x1s1kP(Z=k)+s1xP(Zx)=1α(s11)[s1α/(1+α)]x1α(s11).\displaystyle=\sum_{k=0}^{x-1}s_{1}^{k}\operatorname{P}\,(Z=k)+s_{1}^{x}\operatorname{P}\,(Z\geq x)=\frac{1-\alpha(s_{1}-1)\left[s_{1}\alpha/(1+\alpha)\right]^{x}}{1-\alpha(s_{1}-1)}. (13)

Therefore,

ΨXt,Xt1(s1,s2)\displaystyle\Psi_{X_{t},X_{t-1}}(s_{1},s_{2}) =Ψϵt(s1)E(s2X1α(s11)α(s11)1α(s11)(s1s2α1+α)X)\displaystyle=\Psi_{\epsilon_{t}}(s_{1})\operatorname{E}\,\left(\frac{s_{2}^{X}}{1-\alpha(s_{1}-1)}-\frac{\alpha(s_{1}-1)}{1-\alpha(s_{1}-1)}\left(\frac{s_{1}s_{2}\alpha}{1+\alpha}\right)^{X}\right)
=Ψϵt(s1)1α(s11)[ΨX(s2)α(s11)ΨX(s1s2α1+α)].\displaystyle=\frac{\Psi_{\epsilon_{t}}(s_{1})}{1-\alpha(s_{1}-1)}\left[\Psi_{X}(s_{2})-\alpha(s_{1}-1)\Psi_{X}\left(\frac{s_{1}s_{2}\alpha}{1+\alpha}\right)\right].

Proposition 8.

The 1-step ahead conditional mean and conditional variance are given by

E(Xt|Xt1)\displaystyle\operatorname{E}\,(X_{t}\,|\,X_{t-1}) =α[1(α1+α)Xt1]+μϵ,\displaystyle=\alpha\left[1-\left(\displaystyle\frac{\alpha}{1+\alpha}\right)^{X_{t-1}}\right]+\mu_{\epsilon},
Var(Xt|Xt1)\displaystyle\operatorname{Var}\,(X_{t}\,|\,X_{t-1}) =α[1(α1+α)Xt1][1+α(1+(α1+α)Xt1)]2αXt1(α1+α)Xt1+σϵ2,\displaystyle=\alpha\left[1-\left(\frac{\alpha}{1+\alpha}\right)^{X_{t-1}}\right]\left[1+\alpha\left(1+\left(\frac{\alpha}{1+\alpha}\right)^{X_{t-1}}\right)\right]-2\alpha X_{t-1}\left(\frac{\alpha}{1+\alpha}\right)^{X_{t-1}}+\sigma_{\epsilon}^{2},

respectively.

Proof.

From the definition of the MGWI process, we obtain that

E(Xt|Xt1=x)=E(αXt1+ϵt|Xt1=x)=E(αXt1|Xt1=x)+μϵ,\operatorname{E}\,(X_{t}\,|\,X_{t-1}=x)=\operatorname{E}\,(\alpha\operatorname{\triangle}X_{t-1}+\epsilon_{t}\,|\,X_{t-1}=x)=\operatorname{E}\,(\alpha\operatorname{\triangle}X_{t-1}\,|\,X_{t-1}=x)+\mu_{\epsilon},

for all x=0,1,x=0,1,\dots. The conditional expectation above can be obtained from Proposition 2 with XX being a degenerate random variable at xx (i.e. P(X=x)=1P(X=x)=1). Then, it follows that

E(Xt|Xt1=x)=α[1(α1+α)x]+μϵ.\operatorname{E}\,(X_{t}\,|\,X_{t-1}=x)=\alpha\left[1-\left(\frac{\alpha}{1+\alpha}\right)^{x}\right]+\mu_{\epsilon}.

The conditional variance can be derived analogously, so details are omitted. ∎

Remark 2.

Note that the conditional expectation and variance given in Proposition 8 are non-linear on Xt1X_{t-1} in contrast with the classic GW/INAR processes where they are linear.

Proposition 9.

The autocovariance and autocorrelation functions at lag 1 of the Geo-MGWI process are respectively given by

γ(1)Cov(Xt,Xt1)=μα(1+μ)(1+α)(1+μ+α)2andρ(1)Corr(Xt,Xt1)=α(1+α)(1+μ+α)2.\gamma(1)\equiv\operatorname{Cov}(X_{t},X_{t-1})=\frac{\mu\alpha(1+\mu)(1+\alpha)}{(1+\mu+\alpha)^{2}}\quad\mbox{and}\quad\rho(1)\equiv\operatorname{Corr}\,(X_{t},X_{t-1})=\frac{\alpha(1+\alpha)}{(1+\mu+\alpha)^{2}}. (14)
Proof.

We have that Cov(Xt,Xt1)=E(XtXt1)E(Xt)E(Xt1)\operatorname{Cov}(X_{t},X_{t-1})=\operatorname{E}\,(X_{t}X_{t-1})-\operatorname{E}\,(X_{t})\operatorname{E}\,(X_{t-1}), with

E(XtXt1)\displaystyle\operatorname{E}\,(X_{t}X_{t-1}) =E[E(XtXt1|Xt1)]=E[Xt1E(Xt|Xt1)]\displaystyle=\operatorname{E}\,\left[\operatorname{E}\,(X_{t}X_{t-1}\,|\,X_{t-1})\right]=\operatorname{E}\,\left[X_{t-1}\operatorname{E}\,(X_{t}\,|\,X_{t-1})\right]
=αE(Xt1)αE[Xt1(α1+α)Xt1]+μϵE(Xt1)\displaystyle=\alpha\operatorname{E}\,(X_{t-1})-\alpha\operatorname{E}\,\left[X_{t-1}\left(\frac{\alpha}{1+\alpha}\right)^{X_{t-1}}\right]+\mu_{\epsilon}\operatorname{E}\,(X_{t-1})
=μαμα2(1+α)(1+μ+α)2+μ2(1+μ)1+μ+α.\displaystyle=\mu\alpha-\frac{\mu\alpha^{2}(1+\alpha)}{(1+\mu+\alpha)^{2}}+\frac{\mu^{2}(1+\mu)}{1+\mu+\alpha}.

After some algebra, the result follows. ∎

In the following proposition, we obtain an expression for the conditional expectation E(Xt|Xtk=)E(X_{t}|X_{t-k}=\ell). This function will be important to find the autocovariance function at lag kk\in\mathbb{N} and to perform prediction and/or forecasting.

Proposition 10.

For α>0\alpha>0, define hj=(1+α)j1(1+α)jαjh_{j}=\frac{(1+\alpha)^{j-1}}{(1+\alpha)^{j}-\alpha^{j}} and gj=α(1+α)j1αj(1+α)jαjg_{j}=\frac{\alpha(1+\alpha)^{j-1}-\alpha^{j}}{(1+\alpha)^{j}-\alpha^{j}}, and the real functions

fj(x)=Ψϵ1(αj1)(hj+gjx),f_{j}(x)=\Psi_{\epsilon_{1}}({\alpha_{*}}^{j-1})\left(h_{j}+g_{j}x\right),

j=2,3,j=2,3,\dots, where αα1+α\alpha_{*}\equiv\frac{\alpha}{1+\alpha} and xx\in\mathbb{R}. Finally, let Hk(x)=f2((fk1(fk(x))))H_{k}(x)=f_{2}(\dots(f_{k-1}(f_{k}(x)))). Then, for all {0}\ell\in\mathbb{N}_{*}\equiv\mathbb{N}\cup\{0\},

E(Xt|Xtk=)=α(1Hk(αk))+μϵ,E(X_{t}|X_{t-k}=\ell)=\alpha\left(1-H_{k}\left({\alpha_{*}}^{k\ell}\right)\right)+\mu_{\epsilon}, (15)

for all integer k2k\geq 2.

Proof.

Let t=σ(X1,,Xt)\mathcal{F}_{t}=\sigma(X_{1},\dots,X_{t}) denote the sigma-field generated by the random variables X1,,XtX_{1},\dots,X_{t}. By the Markov property it is clear that

E(Xt|Xtk)=E(Xt|tk)=E[E(Xt|tk+1)|tk],E(X_{t}|X_{t-k})=E(X_{t}|\mathcal{F}_{t-k})=E[E(X_{t}|\mathcal{F}_{t-k+1})|\mathcal{F}_{t-k}], (16)

for all k1k\geq 1. The proof proceeds by induction on kk. Equation (16) and Proposition 8 give us that

E(Xt|Xt2)\displaystyle E(X_{t}|X_{t-2}) =E[E(Xt|Xt1)|Xt2]=E[α(1αXt1)+μϵ|Xt2]\displaystyle=E[E(X_{t}|X_{t-1})|X_{t-2}]=E\left[\alpha(1-{\alpha_{*}}^{X_{t-1}})+\mu_{\epsilon}|X_{t-2}\right]
=α[1E(ααXt2+ϵt1|Xt2)]+μϵ=α[1Ψϵ1(α)E(ααXt2|Xt2)]+μϵ,\displaystyle=\alpha[1-E(\alpha_{*}^{\alpha\operatorname{\triangle}X_{t-2}+\epsilon_{t-1}}|X_{t-2})]+\mu_{\epsilon}=\alpha[1-\Psi_{\epsilon_{1}}(\alpha_{*})E({\alpha_{*}}^{\alpha\operatorname{\triangle}X_{t-2}}|X_{t-2})]+\mu_{\epsilon},

with α=α1+α\alpha_{*}=\frac{\alpha}{1+\alpha}. Using (13), we obtain that

E(Xt|Xt2=)\displaystyle E(X_{t}|X_{t-2}=\ell) =α[1Ψϵ1(α)E(ααXt2|Xt2=)]+μϵ\displaystyle=\alpha[1-\Psi_{\epsilon_{1}}(\alpha_{*})E({\alpha_{*}}^{\alpha\operatorname{\triangle}X_{t-2}}|X_{t-2}=\ell)]+\mu_{\epsilon}
=α[1Ψϵ1(α)1α(α1)[αα/(1+α)]1α(α1)]+μϵ\displaystyle=\alpha\left[1-\Psi_{\epsilon_{1}}(\alpha_{*})\frac{1-\alpha(\alpha_{*}-1)[\alpha_{*}\alpha/(1+\alpha)]^{\ell}}{1-\alpha(\alpha_{*}-1)}\right]+\mu_{\epsilon}
=α[1Ψϵ1(α)(1+α(1+α)2α2+α(1+α)α2(1+α)2α2α2)]+μϵ=α(1H2(α2))+μϵ.\displaystyle=\alpha\left[1-\Psi_{\epsilon_{1}}(\alpha_{*})\left(\frac{1+\alpha}{(1+\alpha)^{2}-\alpha^{2}}+\frac{\alpha(1+\alpha)-\alpha^{2}}{(1+\alpha)^{2}-\alpha^{2}}{\alpha_{*}}^{2\ell}\right)\right]+\mu_{\epsilon}=\alpha\left(1-H_{2}\left({\alpha_{*}}^{2\ell}\right)\right)+\mu_{\epsilon}.

Assume that (15) is true for k=n1k=n-1. Using (16), we have

E(Xt|Xtn)=E(Xt|tn)=E[E(Xt|t(n1))|tn]=α(1E(Hn1(α(n1)Xt(n1))|Xtn))+μϵ.E(X_{t}|X_{t-n})=E(X_{t}|\mathcal{F}_{t-n})=E[E(X_{t}|\mathcal{F}_{t-(n-1)})|\mathcal{F}_{t-n}]=\alpha\left(1-E\left(H_{n-1}\left({\alpha_{*}}^{(n-1)X_{t-(n-1)}}\right)\middle|X_{t-n}\right)\right)+\mu_{\epsilon}.

From the definition of HnH_{n}, we obtain

E(Hn1(α(n1)Xt(n1))|Xtn=)=f2((fn2(E(fn1(α(n1)Xt(n1))|Xtn=)))).E\left(H_{n-1}\left({\alpha_{*}}^{(n-1)X_{t-(n-1)}}\right)\middle|X_{t-n}=\ell\right)=f_{2}\left(\dots\left(f_{n-2}\left(E\left(f_{n-1}\left({\alpha_{*}}^{(n-1)X_{t-(n-1)}}\right)\middle|X_{t-n}=\ell\right)\right)\right)\right).

Note that

E(fn1(α(n1)Xt(n1))|Xtn=)\displaystyle E\left(f_{n-1}\left({\alpha_{*}}^{(n-1)X_{t-(n-1)}}\right)|X_{t-n}=\ell\right) =hn1+gn1E[α(n1)Xt(n1)|Xtn=]\displaystyle=h_{n-1}+g_{n-1}E\left[{\alpha_{*}}^{(n-1)X_{t-(n-1)}}|X_{t-n}=\ell\right]
=hn1+gn1E[α(n1)(αXtn)|Xtn=]\displaystyle=h_{n-1}+g_{n-1}E\left[{\alpha_{*}}^{(n-1)(\alpha\operatorname{\triangle}X_{t-n})}|X_{t-n}=\ell\right]
=hn1+gn1[Ψϵ1(αn1)1α(αn11)[αn1α/(1+α)]1α(αn11)]\displaystyle=h_{n-1}+g_{n-1}\left[\Psi_{\epsilon_{1}}({\alpha_{*}}^{n-1})\frac{1-\alpha({\alpha_{*}}^{n-1}-1)[{\alpha_{*}}^{n-1}\alpha/(1+\alpha)]^{\ell}}{1-\alpha({\alpha_{*}}^{n-1}-1)}\right]
=hn1+gn1[Ψϵ1(αn1)((1+α)n1(1+α)nαn+α(1+α)n1αn(1+α)nαnαn)]\displaystyle=h_{n-1}+g_{n-1}\left[\Psi_{\epsilon_{1}}({\alpha_{*}}^{n-1})\left(\frac{(1+\alpha)^{n-1}}{(1+\alpha)^{n}-\alpha^{n}}+\frac{\alpha(1+\alpha)^{n-1}-\alpha^{n}}{(1+\alpha)^{n}-\alpha^{n}}{\alpha_{*}}^{n\ell}\right)\right]
=fn1(fn(αn)).\displaystyle=f_{n-1}(f_{n}({\alpha_{*}}^{n\ell})).

Therefore, E(Hn1(α(n1)Xt(n1))|Xtn=)=f2((fn1(fn(αnl))))=Hn(αn)E\left(H_{n-1}\left({\alpha_{*}}^{(n-1)X_{t-(n-1)}}\right)\middle|X_{t-n}=\ell\right)=f_{2}(\dots(f_{n-1}(f_{n}({\alpha_{*}}^{nl}))))=H_{n}({\alpha_{*}}^{n\ell}), and hence we get the desired expression E(Xt|Xtn=)=α(1Hn(αn))+μϵE(X_{t}|X_{t-n}=\ell)=\alpha\left(1-H_{n}\left({\alpha_{*}}^{n\ell}\right)\right)+\mu_{\epsilon}, which completes the proof. ∎

Proposition 11.

Let hjh_{j}, gjg_{j} be as in Proposition 10 and write h~j=μhj\tilde{h}_{j}=\mu h_{j}, for jj\in\mathbb{N}. It holds that

γ(k):=Cov(Xt,Xtk)=αμ[1Hk(G(α,μ,k))]+μ(μϵμ),\gamma(k):=Cov(X_{t},X_{t-k})=\alpha\mu\left[1-H_{k}(G(\alpha,\mu,k))\right]+\mu\left(\mu_{\epsilon}-\mu\right),

where G(α,μ,k)=αk(1+μ(1αk))2G(\alpha,\mu,k)=\frac{\alpha_{*}}{k(1+\mu(1-{\alpha_{*}}^{k}))^{2}}, and Hk()H_{k}(\cdot) as defined in Proposition 10, for kk\in\mathbb{N}.

Proof.

Note that

γ(k)=E(E(XtXtk|Xtk))μ2=E(XtkE(Xt|Xtk))μ2=α(μE(XtkHk(αkXtk)))+μ(μϵμ),\gamma(k)=E(E(X_{t}X_{t-k}|X_{t-k}))-\mu^{2}=E(X_{t-k}E(X_{t}|X_{t-k}))-\mu^{2}=\alpha\left(\mu-E\left(X_{t-k}H_{k}\left({\alpha_{*}}^{kX_{t-k}}\right)\right)\right)+\mu(\mu_{\epsilon}-\mu), (17)

where the third equality follows by (15). A thorough inspection of the definition of HkH_{k} gives

E(XtkHk(αkXtk))=f~2((f~k(E(XtkαkXtk)))),E\left(X_{t-k}H_{k}\left({\alpha_{*}}^{kX_{t-k}}\right)\right)=\tilde{f}_{2}\left(\dots\left(\tilde{f}_{k}\left(E\left(X_{t-k}{\alpha_{*}}^{kX_{t-k}}\right)\right)\right)\right), (18)

where we have defined f~j(x)=Ψϵ1(αj1)(h~j+gjx)\tilde{f}_{j}(x)=\Psi_{\epsilon_{1}}({\alpha_{*}}^{j-1})\left(\tilde{h}_{j}+g_{j}x\right), for jj\in\mathbb{N}.

Note that the argument of the function in (18) is just a constant times the derivative of ΨX1(s)\Psi_{X_{1}}(s) with respect to ss and evaluated at α{\alpha_{*}}. More specifically,

E(XtkαkXtk)=αkΨX1(αk)=μαk(1+μ(1αk))2=μG(α,μ,k).E\left(X_{t-k}{\alpha_{*}}^{kX_{t-k}}\right)=\frac{\alpha_{*}}{k}\Psi^{\prime}_{X_{1}}({\alpha_{*}}^{k})=\mu\frac{\alpha_{*}}{k(1+\mu(1-{\alpha_{*}}^{k}))^{2}}=\mu G(\alpha,\mu,k). (19)

The second equality follows from (4). Plugging (19) in (18), we obtain

E(XtkHk(αkXtk))=f~2((f~k(μG(α,μ,k))))=μf2((fk(G(α,μ,k))))=μHk(G(α,μ,k)).E\left(X_{t-k}H_{k}\left({\alpha_{*}}^{kX_{t-k}}\right)\right)=\tilde{f}_{2}(\dots(\tilde{f}_{k}(\mu G(\alpha,\mu,k))))=\mu f_{2}\left(\dots\left({f}_{k}\left(G\left(\alpha,\mu,k\right)\right)\right)\right)=\mu H_{k}(G(\alpha,\mu,k)). (20)

The result follows by plugging (20) in (17). ∎

4 Parameter estimation

In this section, we discuss estimation procedures for the geometric MGWI process through conditional least squares (CLS) and maximum likelihood methods. We assume that X1,,XnX_{1},\ldots,X_{n} is a trajectory from the Geo-MGWI model with observed values x1,,xnx_{1},\ldots,x_{n}, where nn stands for the sample size. We denote the parameter vector by 𝜽(μ,α)\boldsymbol{\theta}\equiv(\mu,\alpha)^{\top}.

For the CLS method, we define the function Qn(𝜽)Q_{n}(\boldsymbol{\theta}) as

Qn(𝜽)t=2n{xtE(Xt|Xt1=xt1)}2=t=2n{xtα[1(α1+α)xt1]μ(1+μ)1+μ+α}2.Q_{n}(\boldsymbol{\theta})\equiv\sum_{t=2}^{n}\left\{x_{t}-\operatorname{E}\,(X_{t}\,|\,X_{t-1}=x_{t-1})\right\}^{2}=\sum_{t=2}^{n}\left\{x_{t}-\alpha\left[1-\left(\frac{\alpha}{1+\alpha}\right)^{x_{t-1}}\right]-\frac{\mu(1+\mu)}{1+\mu+\alpha}\right\}^{2}. (21)

The CLS estimators are obtained as the argument that minimizes Qn(𝜽)Q_{n}(\boldsymbol{\theta}), i.e.

𝜽^cls=argmin𝜽Qn(𝜽).\hat{\boldsymbol{\theta}}_{cls}=\operatorname*{arg\,min}_{\boldsymbol{\theta}}Q_{n}(\boldsymbol{\theta}). (22)

Since we do not have an explicit expression for 𝜽^cls\hat{\boldsymbol{\theta}}_{cls}, numerical optimization methods are required to solve (22). This can be done through optimizer packages implemented in softwares such as R (R Core Team, 2021) and MATLAB. The gradient function associated with Qn()Q_{n}(\cdot) can be provided for these numerical optimizers and is given by

Qn(𝜽)μ\displaystyle\frac{\partial Q_{n}(\boldsymbol{\theta})}{\partial\mu} =2[1α(1+α)(1+μ+α)2]t=2n[xtα(1(α1+α)xt1)μ(1+μ)1+μ+α]and\displaystyle=-2\left[1-\frac{\alpha\left(1+\alpha\right)}{(1+\mu+\alpha)^{2}}\right]\sum_{t=2}^{n}\left[x_{t}-\alpha\left(1-\left(\frac{\alpha}{1+\alpha}\right)^{x_{t-1}}\right)-\frac{\mu(1+\mu)}{1+\mu+\alpha}\right]\quad\text{and}
Qn(𝜽)α\displaystyle\frac{\partial Q_{n}(\boldsymbol{\theta})}{\partial\alpha} =2t=2n[xtα(1(α1+α)xt1)μ(1+μ)1+μ+α][1(α1+α)xt1(1+xt11+α)μ(1+μ)(1+μ+α)2].\displaystyle=-2\sum_{t=2}^{n}\left[x_{t}-\alpha\left(1-\left(\frac{\alpha}{1+\alpha}\right)^{x_{t-1}}\right)-\frac{\mu(1+\mu)}{1+\mu+\alpha}\right]\left[1-\left(\frac{\alpha}{1+\alpha}\right)^{x_{t-1}}\left(1+\frac{x_{t-1}}{1+\alpha}\right)-\frac{\mu(1+\mu)}{(1+\mu+\alpha)^{2}}\right].

A strategy to get the standard errors of the CLS estimates based on bootstrap is proposed and illustrated in our empirical illustrations; please see Section 6.

We now discuss the maximum likelihood estimation (MLE) method. Note that our proposed Geo-MGWI process is a Markov chain (by definition) and therefore the likelihood function can be expressed in terms of the 1-step transition probabilities derived in Proposition 6. The MLE estimators are obtained as the argument that maximizes the log-likelihood function, that is, 𝜽^mle=argmax𝜽n(𝜽)\hat{\boldsymbol{\theta}}_{mle}=\operatorname*{arg\,max}_{\boldsymbol{\theta}}\ell_{n}(\boldsymbol{\theta}), with

n(𝜽)\displaystyle\ell_{n}(\boldsymbol{\theta}) =t=2nlogP(Xt=xt|Xt1=xt1)+logP(X1=x1),\displaystyle=\sum_{t=2}^{n}\log\operatorname{P}\,(X_{t}=x_{t}\,|\,X_{t-1}=x_{t-1})+\log\operatorname{P}\,(X_{1}=x_{1}), (23)

where the conditional probabilities in (23)\eqref{E:loglik} are given by (11) and P(X1=x1)\operatorname{P}\,(X_{1}=x_{1}) is the probability function of a geometric distribution with mean μ\mu. There is no closed-form expression available for 𝜽^mle\hat{\boldsymbol{\theta}}_{mle}. The maximization of (23) can be accomplished through numerical methods such as the Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm implemented in the R package optim. The standard errors of the maximum likelihood estimates can be obtained by using the Hessian matrix associated with (23), which can be evaluated numerically.

In the remaining of this section, we examine and compare the finite-sample behavior of the CLS and MLE methods via Monte Carlo (MC) simulation with 1000 replications per set of parameter configurations, with the parameter estimates computed under both approaches. All the numerical experiments presented in this paper were carried out using the R programming language.

Refer to caption
Figure 1: Sample paths for the Geo-MGWI process and their respective ACF and PACF under Scenarios I (top row) and IV (bottom row) with n=100n=100.
Table 1: Empirical mean and RMSE (within parentheses) of the parameter estimates based on the MLE and CLS methods for the Geo-MGWI process under the Scenarios I, II, III, and IV, and for sample sizes n=100,200,500,1000n=100,200,500,1000.
    Scenario I: μ=2.0,α=1.0\mu=2.0,\,\alpha=1.0
    nn     μ^mle\hat{\mu}_{mle}     α^mle\hat{\alpha}_{mle}     μ^cls\hat{\mu}_{cls}     α^cls\hat{\alpha}_{cls}
    100     1.996 (0.281)     0.957 (0.425)     1.999 (0.282)     0.962 (0.698)
    200     1.995 (0.197)     1.021 (0.290)     1.998 (0.197)     1.059 (0.536)
    500     2.013 (0.124)     0.987 (0.177)     2.014 (0.125)     0.991 (0.339)
    1000     1.998 (0.088)     0.998 (0.128)     1.998 (0.088)     0.988 (0.238)
    Scenario II: μ=1.2,α=0.5\mu=1.2,\,\alpha=0.5
    nn     μ^mle\hat{\mu}_{mle}     α^mle\hat{\alpha}_{mle}     μ^cls\hat{\mu}_{cls}     α^cls\hat{\alpha}_{cls}
    100     1.206 (0.181)     0.486 (0.289)     1.208 (0.182)     0.556 (0.482)
    200     1.197 (0.128)     0.491 (0.205)     1.198 (0.129)     0.495 (0.327)
    500     1.196 (0.082)     0.498 (0.119)     1.196 (0.082)     0.490 (0.197)
    1000     1.200 (0.058)     0.506 (0.090)     1.200 (0.058)     0.494 (0.143)
    Scenario III: μ=0.5,α=1.5\mu=0.5,\,\alpha=1.5
    nn     μ^mle\hat{\mu}_{mle}     α^mle\hat{\alpha}_{mle}     μ^cls\hat{\mu}_{cls}     α^cls\hat{\alpha}_{cls}
    100     0.499 (0.130)     1.515 (0.523)     0.498 (0.132)     1.487 (0.831)
    200     0.499 (0.091)     1.514 (0.387)     0.498 (0.093)     1.495 (0.595)
    500     0.496 (0.058)     1.490 (0.236)     0.496 (0.059)     1.472 (0.356)
    1000     0.500 (0.042)     1.502 (0.174)     0.500 (0.044)     1.524 (0.299)
    Scenario IV: μ=0.3,α=0.5\mu=0.3,\,\alpha=0.5
    nn     μ^mle\hat{\mu}_{mle}     α^mle\hat{\alpha}_{mle}     μ^cls\hat{\mu}_{cls}     α^cls\hat{\alpha}_{cls}
    100     0.298 (0.078)     0.506 (0.271)     0.298 (0.078)     0.504 (0.340)
    200     0.296 (0.057)     0.491 (0.186)     0.297 (0.057)     0.492 (0.244)
    500     0.299 (0.037)     0.496 (0.120)     0.300 (0.037)     0.504 (0.157)
    1000     0.299 (0.026)     0.499 (0.087)     0.299 (0.026)     0.500 (0.110)

We consider four simulation scenarios with different values for 𝜽=(μ,α)\boldsymbol{\theta}=(\mu,\alpha)^{\top}, namely: (I) 𝜽=(2.0,1.0)\boldsymbol{\theta}=(2.0,1.0)^{\top}, (II) 𝜽=(1.2,0.5)\boldsymbol{\theta}=(1.2,0.5)^{\top}, (III) 𝜽=(0.5,1.5)\boldsymbol{\theta}=(0.5,1.5)^{\top}, and (IV) 𝜽=(0.3,0.5)\boldsymbol{\theta}=(0.3,0.5)^{\top}. To illustrate these configurations, we display in Figure 1 simulated trajectories from the Geo-MGWI process and their associated ACF and PACF under Scenarios I and IV. In Table 1, we report the empirical mean and root mean squared error (RMSE) of the parameter estimates obtained from the MC simulation based on the MLE and CLS methods. We can observe that both approaches produce satisfactory results and also a slight advantage of the MLE estimators over the CLS for estimating α\alpha, mainly in terms of RMSE, which is already expected. This advantage can also be seen from Figure 2, which presents boxplots of the parameter estimates for μ\mu and α\alpha under the Scenarios I and IV with sample sizes n=100n=100 and n=500n=500. In general, the estimation procedures considered here produce estimates with bias and RMSE decreasing towards zero as the sample size increases, therefore giving evidence of consistency.

Refer to caption
Figure 2: Boxplots of the MLE and CLS estimates for the Geo-MGWI process under the Scenarios I (top row) and IV (bottom row), and for sample sizes n=100,500n=100,500.

5 Dealing with non-stationarity

In many practical situations, stationarity can be a non-realistic assumption; for instance, see Brännäs (1995), Enciso-Mora et al. (2009), and Wang (2020) for works that investigate non-stationary Poisson INAR process. Motivated by that, in this section, we propose a non-stationary version of the Geo-MGWI process allowing for time-varying parameters. Consider

μt=exp(𝐰t𝜷)andαt=exp(𝐯t𝜸),\displaystyle\mu_{t}=\exp({\bf w}_{t}^{\top}\boldsymbol{\beta})\quad\mbox{and}\quad\alpha_{t}=\exp({\bf v}_{t}^{\top}\boldsymbol{\gamma}),

where 𝐰t{\bf w}_{t} and 𝐯t{\bf v}_{t} are p×1p\times 1 and q×1q\times 1 covariate vectors for t1t\geq 1, and 𝜷\boldsymbol{\beta} and 𝜸\boldsymbol{\gamma} are p×1p\times 1 and q×1q\times 1 vectors of associated regression coefficients.

We define a time-varying or non-stationary Geo-MGWI process by

Xt=αtXt1+ϵt,t=2,3,,\displaystyle X_{t}=\alpha_{t}\operatorname{\triangle}X_{t-1}+\epsilon_{t},\,\,\,t=2,3,\ldots, (24)

and X1Geo(μ1)X_{1}\sim\mbox{Geo}(\mu_{1}), where αtXt1=min(Xt1,Zt)\alpha_{t}\operatorname{\triangle}X_{t-1}=\min(X_{t-1},Z_{t}), {Zt}t\{Z_{t}\}_{t\in\mathbb{N}} is an independent sequence with ZtGeo(αt)Z_{t}\sim\mbox{Geo}(\alpha_{t}), {ϵt}t1\{\epsilon_{t}\}_{t\geq 1} are independent random variables with ϵtZMG(αt1+αt+μt,μt)\epsilon_{t}\sim\mbox{ZMG}\left(\dfrac{\alpha_{t}}{1+\alpha_{t}+\mu_{t}},\mu_{t}\right), for t2t\geq 2. It is also assumed that ϵt\epsilon_{t} is independent of XtlX_{t-l} and Ztl+1Z_{t-l+1}, for all l1l\geq 1. Under these assumptions, the marginals of the process (24) are Geo(μt)\mbox{Geo}(\mu_{t}) distributed, for tt\in\mathbb{N}.

We consider two estimation methods for the parameter vector 𝜽=(𝜷,𝜸)\boldsymbol{\theta}=(\boldsymbol{\beta},\boldsymbol{\gamma})^{\top}. The first one is based on the conditional least squares. The CLS estimator of 𝜽\boldsymbol{\theta} is obtained by minimizing (21) with μt\mu_{t} and αt\alpha_{t} instead of μ\mu and α\alpha, respectively. According to Wang (2020), this procedure might not be accurate in the sense that non-significant covariates can be included in the model. In that paper, a penalized CLS (PCLS) method is considered. Hence, a more accurate estimator is obtained by minimizing Q~n(𝜽)=Qn(𝜽)+nj=1p+qPδ(|θi|)\widetilde{Q}_{n}(\boldsymbol{\theta})=Q_{n}(\boldsymbol{\theta})+n\sum_{j=1}^{p+q}P_{\delta}(|\theta_{i}|), where Pδ()P_{\delta}(\cdot) is a penalty function and δ\delta is a tuning parameter. See Wang (2020) for possible choices of penalty function. This can be used as a selection criterion and we hope to explore it in a future paper. A second method for estimating the parameters is the maximum likelihood method. The log-likelihood function assumes the form (23) with μ\mu and α\alpha replaced by μt\mu_{t} and αt\alpha_{t}, respectively.

For the non-stationary case, we carry out a second set of Monte Carlo simulations by considering trend and seasonal covariates in the model as follows:

μt=exp(β0+β1t/n+β2cos(2πt/12))andαt=exp(γ0+γ1t/n),\mu_{t}=\exp(\beta_{0}+\beta_{1}t/n+\beta_{2}\cos(2\pi t/12))\quad\mbox{and}\quad\alpha_{t}=\exp(\gamma_{0}+\gamma_{1}t/n),

for t=1,,nt=1,\ldots,n. The above structure aims to mimic realistic situations when dealing with epidemic diseases. We here set the following scenarios: (V) (β0,β1,β2,γ0,γ1)=(2.0,1.0,0.7,2.0,1.0)(\beta_{0},\beta_{1},\beta_{2},\gamma_{0},\gamma_{1})=(2.0,1.0,0.7,2.0,1.0) and (VI) (β0,β1,β2,γ0,γ1)=(3.0,1.0,0.5,3.0,2.0)(\beta_{0},\beta_{1},\beta_{2},\gamma_{0},\gamma_{1})=(3.0,1.0,0.5,3.0,2.0). We consider 500 Monte Carlo replications and the sample sizes n=100,200,500,1000n=100,200,500,1000. Table 2 reports the empirical mean and the RMSE (within parentheses) of the parameter estimates based on the MLE and CLS methods. We can observed that the MLE method outperforms the CLS method for all configurations considered, as expected since we are generating time series data from the “true” model. This can be also seen from Figure 3, which presents the boxplots of MLE and CLS estimates under the Scenarios V with sample sizes n=200,500n=200,500. Regardless, note that the bias and RMSE of the CLS estimates decrease as the sample size increases.

Table 2: Empirical mean and RMSE (within parentheses) of the parameter estimates based on the MLE and CLS methods for the non-stationary Geo-MGWI process under the Scenarios V and VI, and for sample sizes n=100,200,500,1000n=100,200,500,1000.
Scenario V   (β0,β1,β2,γ0,γ1)=(2.0,1.0,0.7,2.0,1.0)(\beta_{0},\beta_{1},\beta_{2},\gamma_{0},\gamma_{1})=(2.0,1.0,0.7,2.0,1.0)
β^0\hat{\beta}_{0} β^1\hat{\beta}_{1} β^2\hat{\beta}_{2} γ^0\hat{\gamma}_{0} γ^1\hat{\gamma}_{1}
n=100n=100 MLE 1.969 (0.278) 1.006 (0.473) 0.662 (0.209) 1.939 (0.635) 1.027 (0.741)
CLS 1.398 (2.425) 1.373 (1.997) 0.914 (1.317) 1.777 (1.649) 0.469 (2.531)
n=200n=200 MLE 1.984 (0.206) 0.987 (0.337) 0.677 (0.138) 1.986 (0.453) 0.963 (0.503)
CLS 1.743 (1.212) 1.117 (1.005) 0.813 (0.725) 1.783 (1.295) 0.732 (1.557)
n=500n=500 MLE 1.993 (0.126) 1.007 (0.209) 0.673 (0.091) 2.005 (0.168) 1.003 (0.272)
CLS 1.923 (0.317) 1.059 (0.376) 0.706 (0.241) 1.864 (0.622) 1.184 (1.000)
n=1000n=1000 MLE 1.996 (0.083) 1.006 (0.136) 0.674 (0.064) 2.012 (0.116) 0.999 (0.200)
CLS 1.929 (0.265) 1.061 (0.284) 0.696 (0.205) 1.869 (0.461) 1.230 (0.665)
Scenario VI   (β0,β1,β2,γ0,γ1)=(3.0,1.0,0.5,3.0,2.0)(\beta_{0},\beta_{1},\beta_{2},\gamma_{0},\gamma_{1})=(3.0,1.0,0.5,3.0,2.0)
β^0\hat{\beta}_{0} β^1\hat{\beta}_{1} β^2\hat{\beta}_{2} γ^0\hat{\gamma}_{0} γ^1\hat{\gamma}_{1}
n=100n=100 MLE 2.967 (0.296) 0.965 (0.554) 0.506 (0.221) 2.935 (0.376) 2.049 (0.629)
CLS 2.283 (1.732) 1.264 (1.960) 0.704 (1.098) 3.231 (1.290) 1.291 (2.048)
n=200n=200 MLE 2.981 (0.198) 0.995 (0.363) 0.484 (0.152) 2.996 (0.249) 1.998 (0.435)
CLS 2.357 (1.390) 1.310 (1.410) 0.601 (0.884) 3.561 (1.165) 1.106 (1.855)
n=500n=500 MLE 2.996 (0.121) 0.988 (0.226) 0.484 (0.093) 3.008 (0.147) 1.980 (0.264)
CLS 2.570 (0.950) 1.279 (0.912) 0.641 (0.526) 3.370 (0.886) 1.481 (1.354)
n=1000n=1000 MLE 2.998 (0.083) 1.004 (0.156) 0.477 (0.067) 2.999 (0.099) 2.009 (0.184)
CLS 2.697 (0.623) 1.192 (0.597) 0.601 (0.452) 3.244 (0.714) 1.737 (1.026)
Refer to caption
Figure 3: Boxplots of MLE and CLS estimates for the non-stationary Geo-MGWI process under the Scenarios V with sample sizes n=200n=200 (top row) and n=500n=500 (bottom row).

6 Real data applications

In this section, we discuss the usefulness of our methodology under stationary and non-stationary conditions. In the first empirical example, we consider the monthly number of polio cases reported to the U.S. Centers for Disease Control and Prevention from January 1970 to December 1983, with 168 observations. The data were obtained through the gamlss package in R. Polio (or poliomyelitis) is a disease caused by poliovirus. Symptoms associated with polio can vary from mild flu-like symptoms to paralysis and possibly death, mainly affecting children under 5 years of age. The second example concerns the monthly number of Hansen’s disease cases in the state of Paraíba, Brazil, reported by DATASUS - Information Technology Department of the Brazilian Public Health Care System (SUS), from January 2001 to December 2020, totalizing 240 observations. Hansen’s disease (or leprosy) is a curable infectious disease that is caused by M. leprae. It mainly affects the skin, the peripheral nerves mucosa of the upper respiratory tract, and the eyes. According to the World Health Organization, about 208000 people worldwide are infected with Hansen’s disease. The data are displayed in Table 3.

Table 3: The monthly cases of Hansen’s disease in the state of Paraíba, Brazil.
Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec
2001 60 58 91 72 94 52 54 78 64 111 81 70
2002 55 72 71 70 61 51 80 82 97 107 142 81
2003 92 106 126 78 86 69 91 91 64 83 83 55
2004 67 82 121 84 102 77 83 102 77 59 86 67
2005 59 86 84 102 75 57 82 126 107 123 138 94
2006 88 78 105 91 106 68 85 106 95 80 101 67
2007 78 81 96 68 94 67 66 88 71 84 74 64
2008 79 75 66 81 74 45 82 91 85 74 77 61
2009 53 79 105 81 68 67 64 73 75 76 85 48
2010 51 74 94 64 60 51 54 70 69 68 64 43
2011 66 67 83 77 71 67 58 90 73 59 78 72
2012 71 82 80 64 82 60 83 77 76 60 49 52
2013 54 53 80 83 52 52 79 61 71 61 78 47
2014 61 79 51 63 51 45 61 63 83 63 60 40
2015 64 53 79 43 55 47 48 66 48 48 46 48
2016 39 43 54 34 50 38 38 67 35 44 48 41
2017 40 46 54 43 43 53 45 68 65 44 58 47
2018 64 42 72 62 51 42 43 64 47 48 76 40
2019 63 70 56 54 59 51 60 65 80 85 65 49
2020 57 62 61 16 21 19 35 25 60 63 51 30
2021 35 53 56 41 44 41 32 33 17 5 5 5

6.1 Polio data analysis

We begin the analysis of the polio data by providing plots of the observed time series and the corresponding sample ACF and PACF plots in Figure 4. These plots give us evidence that the count time series is stationary. Table 4 provides a summary of the polio data with descriptive statistics, including mean, median, variance, skewness, and kurtosis. From the results in Table 4, we can observe that counts vary between 0 and 14, with the sample mean and variance equal to 1.333 and 3.505, respectively, which suggests overdispersion of the data.

Refer to caption
Figure 4: Polio data (top panel) and corresponding autocorrelation function (bottom left panel) and partial autocorrelation function (bottom right panel).
Table 4: Descriptive statistics of the polio data.
Minimum Maximum Mean Median Variance Skewness Kurtosis
0 14 1.333 1.000 3.505 3.052 16.818

For comparison purposes, we consider the classic first-order GWI/INAR process with E(Xt|Xt1)=αXt1+μ(1α)E(X_{t}|X_{t-1})=\alpha X_{t-1}+\mu(1-\alpha). This linear conditional expectation on Xt1X_{t-1} holds for the classic stationary INAR processes such as the binomial thinning-based ones, in particular, the Poisson INAR(1) model by Alzaid and Al-Osh (1987). The aim is to evaluate the effect of the nonlinearity of our proposed models on the prediction in comparison to the classic GWI/INAR(1) processes.

We consider the CLS estimation procedure, where just the conditional expectation is considered. This allows for a more flexible approach since no further assumptions are required. To obtain the standard errors of the CLS estimates, we consider a parametric bootstrap where some model satisfying the specific form for the conditional expectation holds. In this first application, for our MGWI process, we consider the geometric model derived in Section 3. For the classic INAR, the Poisson model by Alzaid and Al-Osh (1987) is considered in the bootstrap approach. This strategy to get standard errors has been considered, for example, by Maia et al. (2021) for a class of semiparametric time series models driven by a latent factor. In order to compare the predictive performance of the competing models, we compute the sum of squared prediction errors (SSPE) defined by SSPE=t=2n(xtμ^t)2\text{SSPE}=\sum_{t=2}^{n}(x_{t}-\hat{\mu}_{t})^{2}, where μ^t=E^(Xt|Xt1)\hat{\mu}_{t}=\widehat{E}(X_{t}|X_{t-1}) is the predicted mean at time tt, for t=2,,nt=2,\ldots,n. Table 5 summarizes the fitted models by providing CLS estimates and their respective standard errors, and the SSPE values. The SSPE results in Table 5 show the superior performance of the MGWI process over the classic GWI/INAR process in terms of prediction. This can also be observed from Figure 5, where the MGWI process shows a better agreement between the observed and predicted values.

Table 5: CLS estimates, standard errors, and SSPE values of the fitted MGWI and classic GWI/INAR model for the monthly cases of polio.
Models Parameters Estimates Stand. Errors SSPE
MGWI μ\mu 1.3585 0.2047 522.8987
α\alpha 2.6514 1.2230
GWI/INAR μ\mu 1.3572 0.1627 530.6749
α\alpha 0.3063 0.0772
Refer to caption
Figure 5: Plots of polio data (solid lines) and fitted conditional means (dots) based on the MGWI process (to the left) and classic GWI/INAR process (to the right).

To evaluate the adequacy of our proposed MGWI process, we consider the Pearson residuals defined by Rt(Xtμ^t)/σ^tR_{t}\equiv(X_{t}-\hat{\mu}_{t})/\hat{\sigma}_{t}, where σ^t=Var^(Xt|Xt1)\hat{\sigma}_{t}=\sqrt{\widehat{\mbox{Var}}(X_{t}|X_{t-1})}, for t=2,,nt=2,\ldots,n, where we assume that the conditional variance takes the form given in Proposition 8. Figure 6 presents the Pearson residuals against the time, its ACF, and the qq-plot against the normal quantiles. These plots show that the data correlation was well-captured. On the other hand, the qq-plot suggests that the Pearson residuals are not normally distributed. Actually, this discrepancy is not unusual especially when dealing with low counts; for instance, see Zhu (2011) and Silva and Barreto-Souza (2019). As an alternative way to check the adequacy, we use the normal pseudo-residuals introduced by Dunn and Smyth (1996), which is defined by Rt=Φ1(Ut)R^{\ast}_{t}=\Phi^{-1}(U_{t}), where Φ()\Phi(\cdot) is the standard normal distribution function and UtU_{t} is uniformly distributed on the interval (F𝜽^(xt1),F𝜽^(xt))(F_{\boldsymbol{\hat{\theta}}}(x_{t}-1),F_{\boldsymbol{\hat{\theta}}}(x_{t})), where F𝜽^()F_{\boldsymbol{\hat{\theta}}}(\cdot) is the fitted predictive cumulative distribution function of the MGWI process. Figure 7 shows the pseudo residuals against the time, its ACF, and qq-plot. We can observe that the pseudo-residuals are not correlated and are approximately normally distributed. Therefore, we conclude that the MGWI process provides an adequate fit to the polio count time series data.

Refer to caption
Figure 6: Pearson residuals for the MGWI process fitted to the polio data: residuals against time (top panel), ACF (bottom left panel) and qq-plot (bottom right panel).
Refer to caption
Figure 7: Pseudo-residuals for the MGWI process fitted to the polio data: residuals against time (top panel), ACF (bottom left panel) and qq-plot (bottom right panel).

6.2 Hansen’s disease data analysis

We now analyze Hansen’s disease data. A descriptive data analysis is provided in Table 6. Figure 8 presents the Hansen’s count data and its corresponding sample ACF and PACF plots. This figure provides evidence that the count time series is non-stationary. In particular, we can observe a negative trend. This motivates us to use non-stationarity approaches to handle this data. We consider our non-stationary MGWI process with conditional mean

E(Xt|Xt1)=αt[1(αt1+αt)Xt1]+μt(1+μt)1+μt+αt,\displaystyle E(X_{t}|X_{t-1})=\alpha_{t}\left[1-\left(\dfrac{\alpha_{t}}{1+\alpha_{t}}\right)^{X_{t-1}}\right]+\dfrac{\mu_{t}(1+\mu_{t})}{1+\mu_{t}+\alpha_{t}}, (25)

where the following regression structure is assumed:

μt=exp(β0+β1t/252)andαt=exp(γ0+γ1t/252),fort=1,,252,\displaystyle\mu_{t}=\exp\left(\beta_{0}+\beta_{1}t/252\right)\quad\text{and}\quad\alpha_{t}=\exp\left(\gamma_{0}+\gamma_{1}t/252\right),\,\,\,\text{for}\,\,\,t=1,\ldots,252,

with the term t/252t/252 being a linear trend. For comparison purposes, we also consider the Poisson INAR(1) process allowing for covariates (Brännäs, 1995) with conditional expectation E(Xt|Xt1)=αtXt1+μt(1αt)E(X_{t}|X_{t-1})=\alpha_{t}X_{t-1}+\mu_{t}(1-\alpha_{t}), where

μt=exp(β0+β1t/252)andαt=exp(γ0+γ1t/252)1+exp(γ0+γ1t/252),fort=1,,252.\displaystyle\mu_{t}=\exp\left(\beta_{0}+\beta_{1}t/252\right)\quad\text{and}\quad\alpha_{t}=\dfrac{\exp\left(\gamma_{0}+\gamma_{1}t/252\right)}{1+\exp\left(\gamma_{0}+\gamma_{1}t/252\right)},\,\,\,\text{for}\,\,\,t=1,\ldots,252.
Table 6: Descriptive statistics of the Hansen’s disease data.
Minimum Maximum Mean Median Variance Skewness Kurtosis
5 142 66.63 66 481.103 0.250 3.937
Refer to caption
Figure 8: Hansen’s disease data (top panel) and corresponding autocorrelation function (bottom left panel) and partial autocorrelation function (bottom right panel).

We consider the CLS estimation procedure for both approaches considered here. Table 7 gives us the parameter estimates under the MGWI and PINAR(1) processes, standard errors obtained via bootstrap, and the SSPE values. To get the standard errors for the parameter estimates, we proceed similarly as done in the first application with a slight difference. Since here the counts are high, the geometric assumption cannot be valid. Therefore, we consider a non-stationary MGWI process with innovations following a Poisson distribution with mean μt(1+μt)1+μt+αt\dfrac{\mu_{t}(1+\mu_{t})}{1+\mu_{t}+\alpha_{t}} in our bootstrap scheme. This ensures that the conditional mean is the same as in (25). From Table 7, we have that the trend is significant (using, for example, a significance level at 5%) to explain the marginal mean μt\mu_{t}, but not for the parameter αt\alpha_{t}, under the MGWI model. Furthermore, we note that the sign of the estimate of β\beta is negative, which is in agreement with the observed negative trend. We highlight that the parameter μt\mu_{t} also appears in the autocorrelation structure under our approach, therefore the trend is also significant to explain the autocorrelation of the MGWI process. By looking at the results from the PINAR fitting, we see that the trend is significant to explain αt\alpha_{t} (parameter related to the autocorrelation) but not the marginal mean μt\mu_{t}. Once again, we have that the model producing the smallest SSPE is the MGWI process. So, our proposed methodology is performing better than the classic PINAR model in terms of prediction. The predictive values according to both models along with the observed counts are exhibited in Figure 9.

Table 7: CLS estimates, standard errors, and SSPE values of the fitted non-stationary MGWI and classic PINAR processes for the Hansen’s disease data.
Models Parameters Estimates Stand. Errors SSPE
MGWI β0\beta_{0} 4.3538 0.0671 58742.31
β1\beta_{1} 0.7243-0.7243 0.1178
γ0\gamma_{0} 4.5297 0.6303
γ1\gamma_{1} 0.5613-0.5613 0.9399
PINAR β0\beta_{0} 0.7668-0.7668 0.5332 59919.40
β1\beta_{1} 0.7997 0.8876
γ0\gamma_{0} 4.5290 0.0212
γ1\gamma_{1} 0.6883-0.6883 0.0433
Refer to caption
Figure 9: Plots of Hansen’s disease data (solid line) and fitted conditional means (dots) based on the non-stationary MGWI process (to the left) and PINAR process (to the right).

We now conclude this data analysis by checking if the non-stationary MGWI process fits well the data. Figure 10 provides the Pearson residuals against time, its ACF plot, and the qq-plot of the residuals. By looking at this figure, we have evidence of the adequacy of the MGWI process to fit Hansen’s disease data.

Refer to caption
Figure 10: Pearson residuals for the MGWI process fitted to the Hansen’s disease data: residuals against time (top panel), ACF (bottom left panel) and qq-plot(bottom right panel).

7 Generalization

In this section, we provide an extension of the geometric thinning operator and propose a modified GWI process based on such generalization. As we will see, alternative distributions rather than geometric for the operation in (5) can provide flexible approaches for dealing with different features on count time series. We also discuss how to handle zero-inflation or zero-deflation with respect to the geometric model.

Definition 4.

(Zero-modified geometric (ZMG) thinning operator) Assume that XX is a non-negative integer-valued random variable, independent of Z(η,α)ZMG(1η,α)Z^{(\eta,\alpha)}\sim\mathrm{ZMG}(1-\eta,\alpha), with α>0\alpha>0 and 1η(1/α,1)1-\eta\in(-1/\alpha,1). We define the zero-modified geometric thinning operator (η,α)(\eta,\alpha)\operatorname{\triangle} by

(η,α)X=dmin(X,Z(η,α)).(\eta,\alpha)\operatorname{\triangle}X\stackrel{{\scriptstyle d}}{{=}}\min\left(X,Z^{(\eta,\alpha)}\right). (26)
Remark 3.

Note that the ZMG operator given in (26) has the geometric thinning operator as a special case when η=1\eta=1 since Z(1,α)Geo(α)Z^{(1,\alpha)}\sim\mathrm{Geo}(\alpha). Further, we stress that the parameterization of the ZMG distribution in terms of 1η1-\eta instead of η\eta will be convenient in what follows. Also, we will omit the dependence of ZZ on (η,α)(\eta,\alpha) to simplify the notation.

Based on the ZMG operator, we can define a modified GWI process {Xt}t\{X_{t}\}_{t\in\mathbb{N}} (similarly as done in Section 3) by

Xt=(η,α)Xt1+ϵt,t,X_{t}=(\eta,\alpha)\operatorname{\triangle}X_{t-1}+\epsilon_{t},\quad t\in\mathbb{N}, (27)

where (η,α)Xt1=min(Xt1,Zt)(\eta,\alpha)\operatorname{\triangle}X_{t-1}=\min\left(X_{t-1},Z_{t}\right), with {Zt}tiidZMG(1η,α)\{Z_{t}\}_{t\in\mathbb{N}}\stackrel{{\scriptstyle iid}}{{\sim}}\mbox{ZMG}(1-\eta,\alpha), {ϵt}t1\{\epsilon_{t}\}_{t\geq 1} is a sequence of iid non-negative integer-valued random variables, called innovations, with ϵt\epsilon_{t} independent of XtlX_{t-l} and Ztl+1Z_{t-l+1}, for all l1l\geq 1, with X0X_{0} being some starting value/random variable. This is basically the same idea as before; we are just replacing the geometric assumption by the zero-modified geometric law in the thinning operation.

We now show that it is possible to construct a stationary Markov chain satisfying (27) and having marginals ZMG-distributed; this could be seen as an alternative model to the zero-modified geometric INAR(1) process proposed by Barreto-Souza (2015). Furthermore, we argue that such construction is not possible under the geometric thinning operator defined in Section 2 (see Remark 4 below), which motivates the ZMG thinning introduced here.

Let XZMG(1π,μ)X{\sim}\mbox{ZMG}(1-\pi,\mu) with μ>0\mu>0 and 1π(1/μ,1)1-\pi\in(-1/\mu,1). For z=0,1,z=0,1,\dots, it holds that

P((η,α)X>z)\displaystyle\operatorname{P}\,((\eta,\alpha)\operatorname{\triangle}X>z) =P(X>z)P(Z(η,α)>z)=πη[(μ1+μ)(α1+α)]z+1.\displaystyle=\operatorname{P}\,(X>z)\operatorname{P}\,(Z^{(\eta,\alpha)}>z)=\pi\eta\left[\left(\frac{\mu}{1+\mu}\right)\left(\frac{\alpha}{1+\alpha}\right)\right]^{z+1}.

In other words, (η,α)XZMG(1ηπ,μα1+μ+α)(\eta,\alpha)\operatorname{\triangle}X\sim\mathrm{ZMG}\left(1-\eta\pi,\frac{\mu\alpha}{1+\mu+\alpha}\right). Writing Ψϵ(s)ΨX(s)Ψ(η,α)X(s)\Psi_{\epsilon}(s)\equiv\dfrac{\Psi_{X}(s)}{\Psi_{(\eta,\alpha)\operatorname{\triangle}X}(s)}, we obtain

Ψϵ(s)\displaystyle\Psi_{\epsilon}(s) =\displaystyle= {1+(1π)μ(1s)1+μ(1s)}/{1+(1πη)μα1+μ+α(1s)1+μα1+μ+α(1s)}\displaystyle\left\{\dfrac{1+(1-\pi)\mu(1-s)}{1+\mu(1-s)}\right\}\Bigg{/}\left\{\dfrac{1+(1-\pi\eta)\frac{\mu\alpha}{1+\mu+\alpha}(1-s)}{1+\frac{\mu\alpha}{1+\mu+\alpha}(1-s)}\right\} (28)
=\displaystyle= {1+(1π)μ(1s)1+(1πη)μα1+μ+α(1s)}{1+μα1+μ+α(1s)1+μ(1s)}φ1(s)φ2(s),\displaystyle\left\{\dfrac{1+(1-\pi)\mu(1-s)}{1+(1-\pi\eta)\frac{\mu\alpha}{1+\mu+\alpha}(1-s)}\right\}\left\{\dfrac{1+\frac{\mu\alpha}{1+\mu+\alpha}(1-s)}{1+\mu(1-s)}\right\}\equiv\varphi_{1}(s)\varphi_{2}(s),

for all ss such that |s|<1+min(μ1,α1)|s|<1+\min(\mu^{-1},\alpha^{-1}), where φ2()\varphi_{2}(\cdot) denotes the pgf of a ZMG(α1+μ+α,μ)\mathrm{ZMG}\left(\frac{\alpha}{1+\mu+\alpha},\mu\right) distribution. In addition to the restrictions on π\pi and η\eta above, assume that πη<1\pi\eta<1, η1\eta\neq 1, and 1π1πη(1+1+μα)<1\frac{1-\pi}{1-\pi\eta}\left(1+\frac{1+\mu}{\alpha}\right)<1. Under these conditions, φ1()\varphi_{1}(\cdot) is the pgf of a ZMG(1π1πη(1+1+μα),(1πη)μα1+μ+α)\mathrm{ZMG}\left(\frac{1-\pi}{1-\pi\eta}\left(1+\frac{1+\mu}{\alpha}\right),(1-\pi\eta)\frac{\mu\alpha}{1+\mu+\alpha}\right) distribution. This implies that Ψϵ()\Psi_{\epsilon}(\cdot) is a proper pgf associated to a convolution between two independent ZMG random variables. Hence, we are able to introduce a MGWI process with ZMG marginals as follows.

Definition 5.

A stationary MGWI process {Xt}t\{X_{t}\}_{t\in\mathbb{N}} with ZMG(1π,μ)\mbox{ZMG}(1-\pi,\mu) marginals (ZMG-MGWI) is defined by assuming that (27) holds with {ϵt}t1\{\epsilon_{t}\}_{t\geq 1} being an iid sequence of random variables with pgf given by (28), and X0ZMG(1π,μ)X_{0}{\sim}\mbox{ZMG}(1-\pi,\mu), with μ>0\mu>0 and 1π(1/μ,1)1-\pi\in(-1/\mu,1).

Remark 4.

Note that we are excluding the case η=1\eta=1 (which corresponds to the geometric thinning operator) since the required inequality 1π1πη(1+1+μα)<1\frac{1-\pi}{1-\pi\eta}\left(1+\frac{1+\mu}{\alpha}\right)<1 does not hold in this case (1+1+μα>11+\frac{1+\mu}{\alpha}>1). This shows that an MGWI process with ZMG marginals cannot be constructed based on the geometric thinning operator defined previously and therefore motivates the ZMG operator.

Acknowledgments

W. Barreto-Souza would like to acknowledge support from KAUST Research Fund. Roger Silva was partially supported by FAPEMIG, grant APQ-00774-21.

References

  • Aleksić and Ristić (2021) Aleksić, M.S., Ristić, M.M. (2021). A geometric minification integer-valued autoregressive model. Applied Mathematical Modelling. 90, 265–280.
  • Alzaid and Al-Osh (1987) Alzaid, A.A., Al-Osh, M.A. (1987). First-order integer-valued autoregressive (INAR(1)) process. Journal of Time Series Analysis. 8, 261–275.
  • Barreto-Souza (2015) Barreto-Souza, W. (2015). Zero-modified geometric INAR(1) process for modelling count time series with deflation or inflation of zeros. Journal of Time Series Analysis. 36, 839–852.
  • Billingsley (1995) Billingsley, P. (1995). Probability and Measure, 3rd edition. Wiley & Sons, New York.
  • Brännäs (1995) Brännäs, K. (1995). Explanatory variables in the AR(1) count data model. Umeå  Economic Studies. 381.
  • Dion et al. (1995) Dion, J.P., Gauthier, G., Latour, A. (1995). Branching processes with immigration and integer-valued time series. Serdica Mathematical Journal. 21, 123–136.
  • Dunn and Smyth (1996) Dunn, P.K., Smyth, G.K. (1996). Randomized quantile residuals. Journal of Computational and Graphical Statistics. 5, 236–244.
  • Enciso-Mora et al. (2009) Enciso-Mora, V., Neal, P., Rao, T.S. (2009). Integer valued AR processes with explanatory variables. Sankhyā – Series B. 71, 248–263.
  • Freeland and McCabe (2005) Freeland, R.M., McCabe, B.P.M. (2005). Asymptotic properties of CLS estimators in the Poisson AR(1) model. Statistics and Probability Letters. 73, 147–153.
  • Kalamkar (1995) Kalamkar, V.A. (1995). Minification processes with discrete marginals. Journal of Applied Probability. 32, 692–706.
  • Littlejohn (1992) Littlejohn, R.P. (1992). Discrete minification processes and reversibility. Journal of Applied Probability. 29, 82–91.
  • Littlejohn (1996) Littlejohn, R.P. (1996). A reversibility relationship for two Markovian time series models with stationary geometric tailed distribution. Stochastic Processes and their Applications. 64, 127–133.
  • Ispány et al. (2003) Ispány, M., Pap, G., Van Zuijlen, M.C.A. (2003). Asymptotic inference for nearly unstable INAR(1) models. Journal of Applied Probability. 40, 750–765.
  • Maia et al. (2021) Maia, G.O., Barreto-Souza, W, Bastos, F.S., Ombao, H. (2021). Semiparametric time series models driven by latent factor. International Journal of Forecasting. 37, 1463–1479.
  • McKenzie (1988) McKenzie, E. (1988). Some ARMA models for dependent sequences of Poisson counts. Advances in Applied Probability. 20, 822–835.
  • R Core Team (2021) R Core Team (2021). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. URL https://www.R-project.org/.
  • Rahimov (2008) Rahimov, I. (2008). Asymptotic distribution of the CLSE in a critical process with immigration. Stochastic Processes and their Applications. 118, 1892–1908.
  • Scotto et al. (2016) Scotto, M.G., Weiß, C.H., Möller, T.A., Gouveia, S. (2016). The max-INAR(1) model for count processes. Test. 27, 850–870.
  • Silva and Barreto-Souza (2019) Silva, R.B., Barreto-Souza, W. (2019). Flexible and robust mixed Poisson INGARCH models. Journal of Time Series Analysis. 40, 788–814.
  • Steutel and van Harn (1979) Steutel, F.W., van Harn, K. (1979). Discrete analogues of self-decomposability and stability. Annals of Probability. 7, 893–899.
  • Wang (2020) Wang, X. (2020). Variable selection for first-order Poisson integer-valued autoregressive model with covariables. Australian and New Zealand Journal of Statistics. 62, 278–295.
  • Wei and Winnicki (1990) Wei, C.Z., Winnicki, J. (1990). Estimation of the mean in the branching process with immigration. Annals of Statistics. 18, 1757–1773.
  • Zhu (2011) Zhu, F. (2011). A negative binomial integer-valued GARCH model. Journal of Time Series Analysis. 32, 54–67.