This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Covert queueing problem with a Markovian statistic

Arti Yardi IIIT Bangalore arti.yardi@iiitb.ac.in    Tejas Bodas\dagger TCS Research, India tejas.bodas@tcs.com
Abstract

Based on the covert communication framework, we consider a covert queueing problem that has a Markovian statistic. Willie jobs arrive according to a Poisson process and require service from server Bob. Bob does not have a queue for jobs to wait and hence when the server is busy, arriving Willie jobs are lost. Willie and Bob enter a contract under which Bob should only serve Willie jobs. As part of the usage statistic, for a sequence of N consecutive jobs that arrived, Bob informs Willie whether each job was served or lost (this is the Markovian statistic). Bob is assumed to be violating the contract and admitting non-Willie (Nillie) jobs according to a Poisson process. For such a setting, we identify the hypothesis testing to be performed (given the Markovian data) by Willie to detect the presence or absence of Nillie jobs. We also characterize the upper bound on arrival rate of Nillie jobs such that the error in the hypothesis testing of Willie is arbitrarily large, ensuring covertness in admitting Nillie jobs.

Index Terms:
Covert communication, Covert queueing, Detection of Markov chains

I Introduction

\daggerThis work was done when the author was a faculty at IIT Dharwad.

In the problem of covert communication, one considers a setup where Alice is transmitting messages to Bob and intruder Willie is snooping over this communication. The aim of Willie is to determine if the communication between Alice and Bob is taking place or not [1, 2]. While this problem has been studied for a wide variety of system models [3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16], in this paper we consider a novel setup of covert queueing problem first studied in [17]. In a covert queueing problem, there are three entities, namely Willie, Bob, and Nillie (for non-Willie). Bob is a server that processes incoming jobs that arrive to its queue. While Bob is obligated to serve only Willie jobs, he may allow some non-Willie or Nillie jobs (for his selfish motives). Willie wants to determine whether Bob is allowing such illegitimate Nillie traffic or not, while Bob wants to have as much Nillie traffic as possible without Willie being able to detect its presence, hence the name covert queueing. Similar to the problem of covert communication, the aim in the covert queueing problem is to determine the asymptotic limit on the arrival rate of Nillie jobs such that Willie is not able to detect the presence of Nillie traffic with high probability.

This covert queueing problem was first introduced in [17] and a recent variant appeared in [18]. In both these models, Willie jobs arrive according the a Poisson process and Willie and Bob enter into an exclusive contract under which Bob should only serve Willie jobs. As proof of exclusive work, Bob periodically informs Willie of some usage statistic. The statistic that Bob uses is typically opaque so that Willie is not able to detect Nillie jobs in an obvious manner. For example, in [17], only the arrival and departure times of Willie jobs are used as the information statistic. Similarly in [18], either the length of successive NN busy periods or the number of jobs served in these busy periods is conveyed. Due to the opaqueness of such statistics, Willie must perform hypothesis testing to infer the presence or absence of Nillie jobs. In turn, Bob desires to admit Nillie jobs in such a way that the probability of error in the hypothesis testing is close to one. The strategy for admitting Nillie jobs is dependent on the information statistic conveyed. For example, [17] employs the probabilistic insert at the end of busy period (IEBP) strategy while in [18], Nillie jobs are admitted according to a Poisson process with rate λb\lambda_{b} and the upper bound of this rate that maintains covertness is determined.

A common feature of both the works described above is that the sequence of information statistic conveyed by Bob are independent and identically distributed (i.i.d.) random variables. In [17], the arrival and departure time information is used to generate a sequence of reconstructed service times that are i.i.d. Similarly, in [18], the statistic includes successive busy period information which is again i.i.d. This i.i.d. nature of the statistic makes the hypothesis testing problem more amenable to analysis.

In this work, we depart from i.i.d. statistics and instead consider a Markovian statistic. We assume that the server Bob does not have a buffer for arriving jobs to wait and the jobs that find server to be busy are lost. (Analysis for case with buffer is for future work.) We assume that Willie and Nillie jobs arrive according to a Poisson process with rates λw\lambda_{w} and λb\lambda_{b} respectively and their service times are i.i.d. with exponential distribution of parameter μ\mu where μ>λw+λb\mu>\lambda_{w}+\lambda_{b}. As part of usage statistics, Bob provides Willie with a sequence of NN random variables X1,,XNX_{1},\ldots,X_{N} (associated with NN successive arrivals), where Xj=1X_{j}=1 if the jj-th arrival finds the server to be busy and is lost and Xj=0X_{j}=0 if the server is idle. Here the NN jobs may include Nillie jobs as well and since the arrival or departure time information is not conveyed, it is difficult for Willie to determine presence or absence of Nillie jobs by observing X1,,XNX_{1},\ldots,X_{N}. This statistic is Markovian and therefore the hypothesis testing as performed by Willie is based on detection of Markov chains with known parameters [19, Ch. 12]. As part of our main result, we formulate this covert queueing setup as a hypothesis testing problem with the Markovian statistic and identify the upper bound on λb\lambda_{b} that ensures covertness. We first obtain a closed form expression for the error exponent associated with this hypothesis testing (see Proposition 1). Using this we then show that the arrival rate λb\lambda_{b} of Nillie jobs should be of order 𝒪(log(K(N))/N)\mathcal{O}(\sqrt{\log(K(N))/N}), where K(N)K(N) is a function of NN that decreases with NN at a sub-exponential rate (see Eq. 5, Theorem 1).

The rest of the paper is organized as follows. In Section II, we first provide details of our problem setup and then recall some of the basics of hypothesis testing between two Markov chains (for the sake of completeness). The main results of the paper are discussed in Section III, followed by some concluding remarks in Section IV.

II System model

Bob is a server that Willie has contracted to serve its jobs. We assume that Willie and Nillie jobs arrive to Bob according to a Poisson process with rates λw\lambda_{w} and λb\lambda_{b} respectively. The server time for each job is exponentially distributed with rate μ\mu. Bob does not have a queue or buffer to store arriving jobs in which case arriving jobs are lost when the server is busy serving a previous arrival. Consider a sequence of NN successive arrivals and let XjX_{j} denote the state of the server (whether it is busy or idle) as seen by the jj-th job. More precisely, Xj=1X_{j}=1 if the arrival found the server to be busy in which case the job is lost. When Xj=0X_{j}=0, the arriving job finds the server to be idle and starts receiving service.

As part of usage statistics, Bob provides Willie with the sequence of random variables X1N{X1,,XN}X_{1}^{N}\coloneqq\{X_{1},\ldots,X_{N}\}, where the choice of NN is arbitrary. Note that Bob does not tell Willie any information about the arrival and departure time of each job. Since the arrivals and Poisson and the service times are exponential, the system can be represented by an M/M/1/1M/M/1/1 queue and the sequence of random variables X1NX_{1}^{N} constitute a Markov chain [20]. Let xjx_{j} be the realization of XjX_{j} for j=1,2,,Nj=1,2,\ldots,N. Given a sequence x1Nx_{1}^{N}, the aim of Willie is to determine if Bob is inserting any non-Willie jobs or not using a binary hypothesis testing problem given by

H0\displaystyle H_{0} :Bob is serving only Willie jobs\displaystyle:\mbox{Bob is serving only Willie jobs} (1)
H1\displaystyle H_{1} :Bob is serving Willie as well as Nillie jobs.\displaystyle:\mbox{Bob is serving Willie as well as Nillie jobs.}

Since the underlying system is an M/M/1/1M/M/1/1 queue, one can see that the value XjX_{j} for the jj-th job only depends on Xj1X_{j-1} but not on any of the variables XkX_{k} for kj2k\leq j-2 [20]. Since XjX_{j} denotes the state of the server as seen by the jj-th arrival and the arrival process is Poisson, {Xj,j0}\{X_{j},j\geq 0\} forms a two state discrete time Markov chain. x1Nx_{1}^{N} denotes an N length realization or sample path of this Markov chain and the transition probability of this Markov chain is a function of the parameters λw,λb,\lambda_{w},\lambda_{b}, and μ\mu. Let PP and QQ denote the state transition matrices under hypotheses H0H_{0} and H1H_{1} respectively.

We assume that both the hypotheses are equally likely. Let PF(N)P_{F}(N) be the probability of rejecting H0H_{0} under the condition it is true and PM(N)P_{M}(N) is the probability of accepting H0H_{0} under the condition H1H_{1} is true. For the equally likely hypotheses, the total error PE(N)P_{E}(N) is equal to PE(N)=(PF(N)+PM(N))/2P_{E}(N)=(P_{F}(N)+P_{M}(N))/2 [21]. In order to achieve covertness, note that Bob wishes to have PE(N)P_{E}(N) close to one. The covertness criteria is defined formally in Section II-B.

II-A Preliminaries about hypothesis testing between two Markov chains [19, Ch. 12]

We shall now summarize some basics about hypothesis testing between two discrete-time, finite-state Markov chains with state transition matrices PP and QQ. Details can be found in [19, Ch. 12] and references therein. For the given sequence x1Nx_{1}^{N}, we wish to determine whether it corresponds to Markov chain with state transition matrix PP (hypothesis H0H_{0}) or QQ (hypothesis H1H_{1}). Let [x1N|Hj]\mathbb{P}[x_{1}^{N}|H_{j}] be the probability of observing x1Nx_{1}^{N} under the condition that hypothesis HjH_{j} is true, for j=0,1j=0,1 and consider the log-likelihood ratio logL(x1N)\log L(x_{1}^{N})111We assume natural logarithm throughout the paper. given by

logL(x1N)log[x1N|H0][x1N|H1].\displaystyle\log L(x_{1}^{N})\coloneqq\log\frac{\mathbb{P}[x_{1}^{N}|H_{0}]}{\mathbb{P}[x_{1}^{N}|H_{1}]}. (2)

The hypothesis testing consists of determining a threshold γ\gamma such that the decision H0H_{0} is true if logL(x1N)γ\log L(x_{1}^{N})\geq\gamma and H1H_{1} is true otherwise. This is represented by

logL(x1N)H1H0γ.\displaystyle\log L(x_{1}^{N})\overset{H_{0}}{\underset{H_{1}}{\gtreqless}}\gamma. (3)

In the asymptotic setting, when NN tends to infinity, one can apply the Gärtner-Ellis theorem to obtain the asymptotic values of PF(N)P_{F}(N) and PM(N)P_{M}(N) which is given by (see Eq. (12.50) of [19])

limN1NlogPF(N)\displaystyle\lim_{N\rightarrow\infty}\frac{1}{N}\log P_{F}(N) =I0(γ)\displaystyle=-I_{0}(\gamma) (4)
limN1NlogPM(N)\displaystyle\lim_{N\rightarrow\infty}\frac{1}{N}\log P_{M}(N) =I1(γ)=γI0(γ),\displaystyle=-I_{1}(\gamma)=\gamma-I_{0}(\gamma),

where I0(γ)I_{0}(\gamma) is termed as the error exponent. It is also called as rate of decay error since it is the rate with which the error in the hypothesis testing goes to zero as NN tends to \infty. Let KF(N)K_{F}(N) and KM(N)K_{M}(N) be some functions of NN that go to zero as NN tends to infinity at a sub-exponential rate, i.e.,

limN1NlogKF(N)=0 & limN1NlogKM(N)=0.\displaystyle\lim_{N\rightarrow\infty}\frac{1}{N}\log K_{F}(N)=0\mbox{ \& }\lim_{N\rightarrow\infty}\frac{1}{N}\log K_{M}(N)=0. (5)

From Eq. 4 and Eq. 5, PF(N)P_{F}(N) and PM(N)P_{M}(N) can be approximated as (see Eq. (12.53) of [19])

PF(N)\displaystyle P_{F}(N) KF(N)exp(I0(γ)N)\displaystyle\approx K_{F}(N)\exp(-I_{0}(\gamma)N) (6)
PM(N)\displaystyle P_{M}(N) KM(N)exp(I1(γ)N).\displaystyle\approx K_{M}(N)\exp(-I_{1}(\gamma)N). (7)

For our problem since both the hypotheses are assumed to be equally likely, we choose the threshold γ\gamma that maximizes the rate of decay of the total error PE(N)=(PF(N)+PM(N))/2P_{E}(N)=(P_{F}(N)+P_{M}(N))/2 and this can be achieved by setting I0(γ)=I1(γ)I_{0}(\gamma)=I_{1}(\gamma). From Eq. 4, this implies that γ=0\gamma=0. In this case, the functions KF(N)K_{F}(N) and KM(N)K_{M}(N) can also be chosen to be equal. Suppose KF(N)=KM(N)=K(N)K_{F}(N)=K_{M}(N)=K(N) and I0(γ)=I1(γ)=IerrI_{0}(\gamma)=I_{1}(\gamma)=I_{err} and using this PE(N)P_{E}(N) can be approximated as

PE(N)K(N)exp(IerrN).\displaystyle P_{E}(N)\approx K(N)\exp(-I_{err}N). (8)

Having outlined the expression for the error in the hypothesis testing, we now define the covertness criteria for our problem. See [1, 3, 8, 18] for a similar definition.

II-B Covertness criterion

Definition 1.

We say that server Bob is able to insert Nillie jobs with ϵ\epsilon-covertness if the error in the hypothesis testing PE(N)P_{E}(N) satisfies PE(N)1ϵP_{E}(N)\geq 1-\epsilon where 0<ϵ<10<\epsilon<1. \square

Note that while the covertness criteria can be defined for arbitrarily chosen values of ϵ\epsilon, typically in the literature ϵ\epsilon is chosen to be close to zero. The results provided in our work are however applicable for arbitrary choice of ϵ\epsilon. In this paper, we are interested in the asymptotic regime, where NN tends to infinity, and hence PE(N)P_{E}(N) is close to K(N)exp(IerrN)K(N)\exp(-I_{err}N) (see Eq. 8). Thus for our problem, for ϵ\epsilon-covertness we need

K(N)exp(IerrN)1ϵ.\displaystyle K(N)\exp(-I_{err}N)\geq 1-\epsilon. (9)

III Main results

In order to characterize the error PE(N)P_{E}(N) in the hypothesis testing of Eq. 1, the key step is to obtain an expression for the error exponent IerrI_{err} (see Eq. 8). In this section, we first provide a closed form expression for this error exponent in Proposition 1. Using this we then study the asymptotic performance of the hypothesis testing, satisfying the given ϵ\epsilon-covertness in Section III-B.

III-A Characterizing the error exponent

Recall that in our setup, the usage statistic provided by Bob is a sequence of random variables X1NX_{1}^{N}, where each XjX_{j} can either be zero or one. Further, X1NX_{1}^{N} form a two state Markov chain either with state transition matrix PP (hypothesis H0H_{0}) or QQ (hypothesis H1H_{1}). The error exponent associated with these two Markov chain hypothesis testing is characterized in the following proposition.

Proposition 1.

When the bufferless server Bob provides Willie with the sequence of server states for NN successive arrivals, the error exponent IerrI_{err} (see Eq. 8) for the hypothesis testing of Eq. 1 is given by

Ierr\displaystyle I_{err} =log(pvq1v+(1p)v(1q)1v) with\displaystyle=-\log\Big{(}p^{v}q^{1-v}+(1-p)^{v}(1-q)^{1-v}\Big{)}\mbox{ with}
v\displaystyle v =log(alogbclog(1/c))/logb,\displaystyle=\log\left(a\frac{\log bc}{\log(1/c)}\right)\Big{/}\log b,

where p=μ/(λw+μ),q=μ/(λb+λw+μ),a=(λb+λw)/μ,b=(λb+λw)/λwp=\mu/(\lambda_{w}+\mu),q=\mu/(\lambda_{b}+\lambda_{w}+\mu),a=(\lambda_{b}+\lambda_{w})/\mu,b=(\lambda_{b}+\lambda_{w})/\lambda_{w} and c=(λw+μ)/(λb+λw+μ)c=(\lambda_{w}+\mu)/(\lambda_{b}+\lambda_{w}+\mu).

Proof.

We first obtain the state transition matrix PP under hypothesis H0H_{0}. Note that the state of the Markov chain (an arrival finds the server either busy or idle), can either be zero or one. Suppose both (j1)(j-1)-th and jj-th arrivals find the server states to be idle, i.e., Xj1=0X_{j-1}=0 and Xj=0X_{j}=0. This happens when the service time for the (j1)(j-1)-th job is less than the interarrival time between the two jobs. Since the inter-arrival times and service times are exponential random variables with parameters μ\mu and λw\lambda_{w} respectively, we have [Xj=0|Xj1=0]=μ/(λw+μ)\mathbb{P}[X_{j}=0|X_{j-1}=0]=\mu/(\lambda_{w}+\mu). The remaining state transition probabilities can be calculated in a similar fashion and matrix PP is given by

P=[μλw+μλwλw+μμλw+μλwλw+μ]=[p1pp1p].\displaystyle P=\begin{bmatrix}\frac{\mu}{\lambda_{w}+\mu}&\frac{\lambda_{w}}{\lambda_{w}+\mu}\\ \frac{\mu}{\lambda_{w}+\mu}&\frac{\lambda_{w}}{\lambda_{w}+\mu}\end{bmatrix}=\begin{bmatrix}p&1-p\\ p&1-p\end{bmatrix}. (10)

Under hypothesis H1H_{1}, the arrival rate of the jobs is λb+λw\lambda_{b}+\lambda_{w} and hence the state transition matrix QQ is given by

Q=[μλw+λb+μλw+λbλw+λb+μμλw+λb+μλw+λbλw+λb+μ]=[q1qq1q].\displaystyle Q=\begin{bmatrix}\frac{\mu}{\lambda_{w}+\lambda_{b}+\mu}&\frac{\lambda_{w}+\lambda_{b}}{\lambda_{w}+\lambda_{b}+\mu}\\ \frac{\mu}{\lambda_{w}+\lambda_{b}+\mu}&\frac{\lambda_{w}+\lambda_{b}}{\lambda_{w}+\lambda_{b}+\mu}\end{bmatrix}=\begin{bmatrix}q&1-q\\ q&1-q\end{bmatrix}. (11)

Corresponding to matrices PP and QQ and for a constant u[0,1]u\in[0,1] we now define a matrix M(u)M(u) as follows

M(u)=[puq1u(1p)u(1q)1upuq1u(1p)u(1q)1u].\displaystyle M(u)=\begin{bmatrix}p^{u}q^{1-u}&(1-p)^{u}(1-q)^{1-u}\\ p^{u}q^{1-u}&(1-p)^{u}(1-q)^{1-u}\end{bmatrix}. (12)

Let r(u)r(u) be the spectral radius of M(u)M(u). For the hypothesis testing between Markov chains with state transition matrices PP and QQ, the error exponent of Eq. 8 is given by [19, Sec.12.2.3]

Ierr=min0u1logr(u).\displaystyle I_{err}=-\min_{0\leq u\leq 1}\log r(u). (13)

To find IerrI_{err} we thus need to first find the spectral radius r(u)r(u) of M(u)M(u). Since p,q0p,q\geq 0, M(u)M(u) is a positive square matrix [22, Definition 2.1] and from Proposition 2.4 of [22], r(u)r(u) is upper and lower bounded by the maximum and minimum row sum of M(u)M(u). Observe that all rows of M(u)M(u) are the same and hence r(u)=puq1u+(1p)u(1q)1ur(u)=p^{u}q^{1-u}+(1-p)^{u}(1-q)^{1-u}. We now substitute the values of pp and qq in r(u)r(u) to obtain

r(u)\displaystyle r(u) =(μλb+λw+μ)(λb+λw+μλw+μ)u+\displaystyle=\left(\frac{\mu}{\lambda_{b}+\lambda_{w}+\mu}\right)\left(\frac{\lambda_{b}+\lambda_{w}+\mu}{\lambda_{w}+\mu}\right)^{u}+ (14)
 (λb+λwλb+λw+μ)(λwλb+λwλb+λw+μλw+μ)u.\displaystyle\mbox{~{}~{}~{}~{}~{}~{}~{}}\left(\frac{\lambda_{b}+\lambda_{w}}{\lambda_{b}+\lambda_{w}+\mu}\right)\left(\frac{\lambda_{w}}{\lambda_{b}+\lambda_{w}}\frac{\lambda_{b}+\lambda_{w}+\mu}{\lambda_{w}+\mu}\right)^{u}.

Suppose Aμ/(λb+λw+μ),B(λb+λw+μ)/(λw+μ),C(λb+λw)/(λb+λw+μ)A\coloneqq\mu/(\lambda_{b}+\lambda_{w}+\mu),B\coloneqq(\lambda_{b}+\lambda_{w}+\mu)/(\lambda_{w}+\mu),C\coloneqq(\lambda_{b}+\lambda_{w})/(\lambda_{b}+\lambda_{w}+\mu) and Dλw(λb+λw+μ)/(λb+λw)(λw+μ)D\coloneqq\lambda_{w}(\lambda_{b}+\lambda_{w}+\mu)/(\lambda_{b}+\lambda_{w})(\lambda_{w}+\mu). Substituting the values of A,B,C,A,B,C, and DD in Eq. 14 we get

r(u)=ABu+CDu.\displaystyle r(u)=AB^{u}+CD^{u}. (15)

To complete the proof, from Eq. 13 we now need to find the value of u[0,1]u\in[0,1] that minimizes logr(u)\log r(u). Let vv be this minimizer. Since log is concave function, vv is given by the solution of ddur(u)=0\frac{\mathrm{d}}{\mathrm{d}u}r(u)=0 and hence using Eq. 15, the minimizer vv should satisfy the following equation

ABvlogB+CDvlogD=0\displaystyle AB^{v}\log B+CD^{v}\log D=0 (16)
(BD)v=Clog(1/D)AlogB\displaystyle\Rightarrow\left(\frac{B}{D}\right)^{v}=\frac{C\log(1/D)}{A\log B} (17)
v=log(CAlog(1/D)log(B))log(B/D).\displaystyle\Rightarrow v=\frac{\log\left(\frac{C}{A}\frac{\log(1/D)}{\log(B)}\right)}{\log(B/D)}. (18)

The required expression of vv is obtained by simplifying Eq. 18 and this completes the proof. ∎

III-B Asymptotic performance of hypothesis testing

For ease of exposition and without loss of generality, we will assume for the remainder of this paper that the service rate μ\mu is set to 11. We now have the following main theorem that characterizes the asymptotic performance of the underlying hypothesis testing problem, while ensuring ϵ\epsilon-covertness.

Theorem 1.

Suppose as the usage statistic, server Bob provides Willie with a sequence of server states for NN successive arrivals. Then Bob can insert Nillie packets covertly if λb8λw(λw+1)2NlogK(N)1ϵ\lambda_{b}\leq\sqrt{\frac{8\lambda_{w}(\lambda_{w}+1)^{2}}{N}\log\frac{K(N)}{1-\epsilon}}, where K(N)K(N) is a function of NN that decreases with NN at a sub-exponential rate (see Eq. 5). Further, λb\lambda_{b} should be of order 𝒪(log(K(N))/N)\mathcal{O}(\sqrt{\log(K(N))/N}).

Proof.

For ϵ\epsilon-covertness we need (see Eq. 9)

K(N)exp(IerrN)1ϵ,\displaystyle K(N)\exp(-I_{err}N)\geq 1-\epsilon, (19)

where the expression of IerrI_{err} is provided in Proposition 1. We observe that for the asymptotic analysis, the expression of IerrI_{err} provided in Proposition 1 is not amenable for further analysis. In the literature, typically such situations are resolved with the aid of Taylor series approximation (for example see [1, 3, 17] and references therein). Along similar lines, we consider the second order Taylor series approximation of IerrI_{err} with respect to λb\lambda_{b} around λb=0\lambda_{b}=0. We denote IerrI_{err} by Ierr(λb)I_{err}(\lambda_{b}) to indicate that it is a function of λb\lambda_{b} (since λw\lambda_{w} is now treated as a constant). We have observed that, the Taylor series expansion of Ierr(λb)I_{err}(\lambda_{b}) is very close to its true value in the asymptotic regime (when NN\rightarrow\infty).

The remainder of this proof is now sub-divided into two parts. We first find the Taylor series expansion of Ierr(λb)I_{err}(\lambda_{b}) and then use it to obtain required asymptotic result for satisfying ϵ\epsilon-covertness criteria. In what follows, all the derivatives are taken with respect to λb\lambda_{b} and f(λb)f^{\prime}(\lambda_{b}) denotes the derivative of a function f(λb)f(\lambda_{b}).

Part-I: Obtaining Taylor series expansion of Ierr(λb)I_{err}(\lambda_{b})

Note that the calculations towards finding the Taylor series expansion are not straightforward (not amenable via MATLAB/Mathematica) and hence we provide these steps in detail. Suppose Ierr(λb)I_{err}(\lambda_{b}) is approximated via the second order Taylor series approximation around λb=0\lambda_{b}=0 as follows

Ierr(λb)Ierr(0)+Ierr(0)λb+Ierr′′(0)2λb2,\displaystyle I_{err}(\lambda_{b})\approx I_{err}(0)+I_{err}^{\prime}(0)\lambda_{b}+\frac{I_{err}^{\prime\prime}(0)}{2}\lambda_{b}^{2}, (20)

where Ierr(0)I_{err}^{\prime}(0) and Ierr′′(0)I_{err}^{\prime\prime}(0) are the first and second derivative of Ierr(λb)I_{err}(\lambda_{b}), evaluated at λb=0\lambda_{b}=0. We have observed numerically that considering terms upto λb2\lambda_{b}^{2} provide a good approximation. In the expression of Ierr(λb)I_{err}(\lambda_{b}), observe that while qq and vv are functions of λb\lambda_{b}, pp is not. To indicate this dependence explicitly we denote qq and vv by q(λb)q(\lambda_{b}) and v(λb)v(\lambda_{b}) respectively. Suppose p¯=1p,q¯(x)=1q(x),F1(λb)=pv(λb)q(λb)1v(λb),F2(λb)=p¯v(λb)q¯(λb)1v(λb)\bar{p}=1-p,\bar{q}(x)=1-q(x),F_{1}(\lambda_{b})=p^{v(\lambda_{b})}q(\lambda_{b})^{1-v(\lambda_{b})},F_{2}(\lambda_{b})=\bar{p}^{v(\lambda_{b})}\bar{q}(\lambda_{b})^{1-v(\lambda_{b})}, and F(λb)=F1(λb)+F2(λb)F(\lambda_{b})=F_{1}(\lambda_{b})+F_{2}(\lambda_{b}). Using this we have Ierr(λb)=logF(λb)I_{err}(\lambda_{b})=-\log F(\lambda_{b}). First and second derivatives of Ierr(λb)I_{err}(\lambda_{b}) are now given by

Ierr(λb)=F(λb)F(λb),Ierr′′(λb)=(F(λb))2F(λb)F′′(λb)(F(λb))2.\displaystyle I_{err}^{\prime}(\lambda_{b})=-\frac{F^{\prime}(\lambda_{b})}{F(\lambda_{b})},I_{err}^{\prime\prime}(\lambda_{b})=\frac{(F^{\prime}(\lambda_{b}))^{2}-F(\lambda_{b})F^{\prime\prime}(\lambda_{b})}{(F(\lambda_{b}))^{2}}. (21)

To obtain the Taylor series expansion, we thus need to evaluate F(λb),F(λb)F(\lambda_{b}),F^{\prime}(\lambda_{b}) and F′′(λb)F^{\prime\prime}(\lambda_{b}) as λb0\lambda_{b}\rightarrow 0. Towards this we first note that limλb0v(λb)=1/2\lim_{\lambda_{b}\rightarrow 0}v(\lambda_{b})=1/2 (the proof for this involves repeated application of L’Hopital’s rule and we skip this due to space constraints). Further, it can be easily verified that q(0)=p,q(0)=p2,q′′(0)=2p3,q¯(0)=1p,q¯(0)=p2q(0)=p,q^{\prime}(0)=-p^{2},q^{\prime\prime}(0)=2p^{3},\bar{q}(0)=1-p,\bar{q}^{\prime}(0)=p^{2}, and q¯′′(0)=2p3\bar{q}^{\prime\prime}(0)=-2p^{3}. We now evaluate F(λb),F(λb)F(\lambda_{b}),F^{\prime}(\lambda_{b}), and F′′(λb)F^{\prime\prime}(\lambda_{b}) as λb0\lambda_{b}\rightarrow 0.

(1) Evaluating limλb0F(λb)\lim_{\lambda_{b}\rightarrow 0}F(\lambda_{b}): F(λb)F(\lambda_{b}) is given by

F(λb)=pv(λb)q(λb)1v(λb)+p¯v(λb)q¯(λb)1v(λb).\displaystyle F(\lambda_{b})=p^{v(\lambda_{b})}q(\lambda_{b})^{1-v(\lambda_{b})}+\bar{p}^{v(\lambda_{b})}\bar{q}(\lambda_{b})^{1-v(\lambda_{b})}. (22)

For any arbitrary functions f(x)f(x) and g(x)g(x), it is known that limxx0f(x)g(x)=f0g0\lim_{x\rightarrow x_{0}}f(x)^{g(x)}=f_{0}^{g_{0}}, where limxx0g(x)=g0\lim_{x\rightarrow x_{0}}g(x)=g_{0} and limxx0f(x)=f0\lim_{x\rightarrow x_{0}}f(x)=f_{0}. It can be seen that limλb0q(λb)=p\lim_{\lambda_{b}\rightarrow 0}q(\lambda_{b})=p and since limλb0v(λb)=1/2\lim_{\lambda_{b}\rightarrow 0}v(\lambda_{b})=1/2 in Eq. 22 we get

limλb0F(λb)=p0.5p(10.5)+p¯0.5p¯(10.5)=1.\displaystyle\lim_{\lambda_{b}\rightarrow 0}F(\lambda_{b})=p^{0.5}p^{(1-0.5)}+\bar{p}^{0.5}\bar{p}^{(1-0.5)}=1. (23)

(2) Evaluating limλb0F(λb)\lim_{\lambda_{b}\rightarrow 0}F^{\prime}(\lambda_{b}): Let us first find F1(λb)F_{1}^{\prime}(\lambda_{b}). With some simple calculations it can be shown that

F1(λb)\displaystyle F_{1}^{\prime}(\lambda_{b}) =pv(λb)q(λb)1v(λb)[v(λb)(logplogq(x))\displaystyle=p^{v(\lambda_{b})}q(\lambda_{b})^{1-v(\lambda_{b})}\bigg{[}v^{\prime}(\lambda_{b})\Big{(}\log p-\log q(x)\Big{)}
 +(1v(λb))q(λb)q(λb)].\displaystyle\mbox{~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}}+\Big{(}1-v(\lambda_{b})\Big{)}\frac{q^{\prime}(\lambda_{b})}{q(\lambda_{b})}\bigg{]}.
F1(λb)T1(λb),\displaystyle\coloneqq F_{1}(\lambda_{b})T_{1}(\lambda_{b}), (24)

where T1(λb)v(λb)(logplogq(x))+(1v(λb))q(λb)q(λb)T_{1}(\lambda_{b})\coloneqq v^{\prime}(\lambda_{b})(\log p-\log q(x))+(1-v(\lambda_{b}))\frac{q^{\prime}(\lambda_{b})}{q(\lambda_{b})}. Substituting q(0)=p,q(0)=p2,q(0)=p,q^{\prime}(0)=-p^{2}, and limλb0v(λb)=1/2\lim_{\lambda_{b}\rightarrow 0}v(\lambda_{b})=1/2 in Eq. 24, it can be verified that limλb0F1(λb)=p,limλb0T1(λb)=p/2\lim_{\lambda_{b}\rightarrow 0}F_{1}(\lambda_{b})=p,\lim_{\lambda_{b}\rightarrow 0}T_{1}(\lambda_{b})=-p/2 and hence limλb0F1(λb)=p2/2\lim_{\lambda_{b}\rightarrow 0}F_{1}^{\prime}(\lambda_{b})=-p^{2}/2. Using similar steps it can be shown that limλb0F2(λb)=p2/2\lim_{\lambda_{b}\rightarrow 0}F_{2}^{\prime}(\lambda_{b})=p^{2}/2 and hence

limλb0F(λb)=limλb0F1(λb)+F2(λb)=p22+p22=0.\displaystyle\lim_{\lambda_{b}\rightarrow 0}F^{\prime}(\lambda_{b})=\lim_{\lambda_{b}\rightarrow 0}F_{1}^{\prime}(\lambda_{b})+F_{2}^{\prime}(\lambda_{b})=\frac{-p^{2}}{2}+\frac{p^{2}}{2}=0. (25)

(3) Evaluating limλb0F′′(λb)\lim_{\lambda_{b}\rightarrow 0}F^{\prime\prime}(\lambda_{b}): Let us first find F1′′(λb)F_{1}^{\prime\prime}(\lambda_{b}). From Eq. 24 we have,

F1′′(λb)=F1(λb)T1(λb)+F1(λb)T1(λb)\displaystyle F_{1}^{\prime\prime}(\lambda_{b})=F_{1}(\lambda_{b})T_{1}^{\prime}(\lambda_{b})+F_{1}^{\prime}(\lambda_{b})T_{1}(\lambda_{b}) (26)

where, with some calculations T1(λb)T_{1}^{\prime}(\lambda_{b}) is obtained as

T1(λb)\displaystyle T_{1}^{\prime}(\lambda_{b}) =2v(λb)q(λb)q(λb)+v′′(λb)(logplogq(λb))+\displaystyle=-2v^{\prime}(\lambda_{b})\frac{q^{\prime}(\lambda_{b})}{q(\lambda_{b})}+v^{\prime\prime}(\lambda_{b})\Big{(}\log p-\log q(\lambda_{b})\Big{)}+
 (1v(λb))[q(λb)q′′(λb)(q(λb))2q(λb)].\displaystyle\mbox{~{}~{}~{}~{}~{}~{}}\Big{(}1-v(\lambda_{b})\Big{)}\bigg{[}\frac{q(\lambda_{b})q^{\prime\prime}(\lambda_{b})-(q^{\prime}(\lambda_{b}))^{2}}{q(\lambda_{b})}\bigg{]}. (27)

Substituting q(0)=p,q(0)=p2,q′′(0)=2p3q(0)=p,q^{\prime}(0)=-p^{2},q^{\prime\prime}(0)=2p^{3} and limλb0v(λb)=1/2\lim_{\lambda_{b}\rightarrow 0}v(\lambda_{b})=1/2 in Eq. 27 we get

limλb0T1(λb)=2pv(0)+p22,\displaystyle\lim_{\lambda_{b}\rightarrow 0}T_{1}^{\prime}(\lambda_{b})=2pv^{\prime}(0)+\frac{p^{2}}{2}, (28)

where v(0)limλb0v(λb)v^{\prime}(0)\coloneqq\lim_{\lambda_{b}\rightarrow 0}v^{\prime}(\lambda_{b}). In the above discussion we showed that, limλb0F1(λb)=p,limλb0F1(λb)=p2/2\lim_{\lambda_{b}\rightarrow 0}F_{1}(\lambda_{b})=p,\lim_{\lambda_{b}\rightarrow 0}F_{1}^{\prime}(\lambda_{b})=-p^{2}/2, and limλb0T1(λb)=p/2\lim_{\lambda_{b}\rightarrow 0}T_{1}(\lambda_{b})=-p/2. Substituting these limits in Eq. 26 and from Eq. 28 we get

limλb0F1′′(λb)=2p2v(0)+p32+p34.\displaystyle\lim_{\lambda_{b}\rightarrow 0}F_{1}^{\prime\prime}(\lambda_{b})=2p^{2}v^{\prime}(0)+\frac{p^{3}}{2}+\frac{p^{3}}{4}. (29)

Using similar steps, limλb0F2′′(λb)\lim_{\lambda_{b}\rightarrow 0}F_{2}^{\prime\prime}(\lambda_{b}) can be evaluated. We skip the details due to space constraints. The limit of F2′′(λb)F_{2}^{\prime\prime}(\lambda_{b}) is given by

limλb0F2′′(λb)=2p2v(0)+p3(p2)2(1p)+p44(1p).\displaystyle\lim_{\lambda_{b}\rightarrow 0}F_{2}^{\prime\prime}(\lambda_{b})=-2p^{2}v^{\prime}(0)+\frac{p^{3}(p-2)}{2(1-p)}+\frac{p^{4}}{4(1-p)}. (30)

Simplifying Eq. 29 and Eq. 30 we get

limλb0F′′(λb)=p34(1p)=14λw(λw+1)2,\displaystyle\lim_{\lambda_{b}\rightarrow 0}F^{\prime\prime}(\lambda_{b})=\frac{-p^{3}}{4(1-p)}=\frac{-1}{4\lambda_{w}(\lambda_{w}+1)^{2}}, (31)

where in the last step we have substituted the value of p=1/(λw+1)p=1/(\lambda_{w}+1) (since μ=1\mu=1). Using Eq. 23, Eq. 25, and Eq. 31 in Eq. 21 we get Ierr(0)=0,Ierr(0)=0,I_{err}(0)=0,I_{err}^{\prime}(0)=0, and Ierr′′(0)=1/(4λw(λw+1)2)I_{err}^{\prime\prime}(0)=1/\big{(}4\lambda_{w}(\lambda_{w}+1)^{2}\big{)}. Substituting this in Eq. 20 we get

Ierr(λb)λb28λw(λw+1)2.\displaystyle I_{err}(\lambda_{b})\approx\frac{\lambda_{b}^{2}}{8\lambda_{w}(\lambda_{w}+1)^{2}}. (32)

Part-II: Covertness criteria

From Eq. 9, for ϵ\epsilon-covertness we need

Ierr1NlogK(N)1ϵ,\displaystyle I_{err}\leq\frac{1}{N}\log\frac{K(N)}{1-\epsilon}, (33)

Substituting Eq. 32 in Eq. 33, we get

λb8λw(λw+1)2NlogK(N)1ϵ\displaystyle\lambda_{b}\leq\sqrt{\frac{8\lambda_{w}(\lambda_{w}+1)^{2}}{N}\log\frac{K(N)}{1-\epsilon}} (34)

and this proves the required upper bound on λb\lambda_{b}. Note that in Eq. 34, λw\lambda_{w} and ϵ\epsilon are constants and K(N)K(N) is a function of NN that decreases with NN at a sub-exponential rate (see Eq. 5). Thus λb\lambda_{b} should be order 𝒪(log(K(N))/N)\mathcal{O}(\sqrt{\log(K(N))/N}) and this completes the proof of the theorem. ∎

IV Conclusion

Departing from the usual i.i.d. statistics, in this work we consider a covert queueing problem with a Markovian statistic. We assume that Bob is a bufferless server who informs Willie about the server states (busy or idle) as seen by N successive arrivals. We formulate this covert queueing setup as a hypothesis testing problem between two Markov Chains PP and QQ and identify the upper bound on λb\lambda_{b} that ensures covertness in admitting Nillie jobs.

As part of future work, it would be interesting to extend this work for the setting where Bob has a queue for the arriving jobs to wait. In this case, the information metric that Bob could use is the sequence of queue lengths as seen by the arriving customers. Relaxing the service times from exponential to general distributions and considering a multi-server setting of the problem is also for future work.

Acknowledgments

This work is supported by the DST-INSPIRE faculty program of Government of India.

References

  • [1] B. Bash, D. Goeckel, and D. Towsley, “Square root law for communication with low probability of detection on AWGN channels,” in Proc. of ISIT, Cambridge, USA, July 2012, pp. 448–452.
  • [2] ——, “Limits of reliable communication with low probability of detection on AWGN channels,” IEEE Journal on Sel. areas in commun. , vol. 31, no. 9, pp. 1921–1930, September 2013.
  • [3] R. Soltani, D. Goeckel, D. Towsley, and A. Houmansadr, “Covert communications on poisson packet channels,” in Proc. of Allerton, UIUC, Illinois, USA, October 2015, pp. 1046–1052.
  • [4] T. Sobers, B. Bash, S. Guha, D. Towsley, and D. Goeckel, “Covert communication in the presence of an uninformed jammer,” IEEE Transactions on Wireless Communications, vol. 16, no. 9, pp. 6193–6206, 2017.
  • [5] K. Huang, H. Wang, D. Towsley, and V. Poor, “LPD communication: A sequential change-point detection perspective,” IEEE Transactions on Communications, vol. 68, no. 4, pp. 2474–2490, 2020.
  • [6] R. Soltani, D. Goeckel, D. Towsley, and A. Houmansadr, “Fundamental limits of covert packet insertion,” IEEE Trans. on Commun. , vol. 68, no. 6, pp. 3401–3414, June 2020.
  • [7] B. Bash, D. Goeckel, and D. Towsley, “Covert communication gains from adversary’s ignorance of transmission time,” IEEE Transactions on Wireless Communications, vol. 15, no. 12, pp. 8394–8405, 2016.
  • [8] P. Che, M. Bakshi, and S. Jaggi, “Reliable deniable communication:hiding messages in noise,” in Proc. of ISIT, Istanbul, Turkey, July 2013, pp. 2945–2949.
  • [9] P. Che, S. Kadhe, M. Bakshi, C. Chan, S. Jaggi, and A. Sprintson, “Reliable, deniable and hidable communication: A quick survey,” in 2014 IEEE Information Theory Workshop, Hobart, Tasmania, Australia, November 2014, pp. 227–231.
  • [10] P. Che, M. Bakshi, C. Chan, and S. Jaggi, “Reliable deniable communication with channel uncertainty,” in 2014 IEEE Information Theory Workshop, Hobart, Tasmania, Australia, November 2014, pp. 30–34.
  • [11] Q. Zhang, M. Bakshi, and S. Jaggi, “Covert communication with polynomial computational complexity,” IEEE Transactions on Information Theory, vol. 66, no. 3, pp. 1354–1384, 2020.
  • [12] K. Arumugam and M. Bloch, “Keyless asynchronous covert communication,” in 2014 IEEE Information Theory Workshop, Cambridge, United Kingdom, September 2016, pp. 191–195.
  • [13] M. Bloch, “Covert communication over noisy channels: A resolvability perspective,” IEEE Trans. on Info. Theory, vol. 62, no. 5, pp. 2334–2354, May 2016.
  • [14] I. Kadampot, M. Tahmasbi, and M. Bloch, “Multilevel-coded pulse-position modulation for covert communications over binary-input discrete memoryless channels,” IEEE Transactions on Information Theory, vol. 66, no. 10, pp. 6001–6023, 2020.
  • [15] M. Tahmasbi, A. Savard, and M. Bloch, “Covert capacity of non-coherent rayleigh-fading channels,” IEEE Transactions on Information Theory, vol. 66, no. 4, pp. 1979–2005, 2020.
  • [16] K. Arumugam and M. Bloch, “Embedding covert information in broadcast communications,” IEEE Transactions on Information Forensics and Security, vol. 14, no. 10, pp. 2787–2801, 2019.
  • [17] B. Jiang, P. Nain, and D. Towsley, “Covert cycle stealing in a single FIFO server,” ACM Transactions on Modeling and Performance Evaluation of Computing Systems, vol. 6, no. 2, June 2021.
  • [18] A. Yardi and T. Bodas, “A covert queueing problem with busy period statistic,” IEEE Communication letters, vol. 25, no. 3, pp. 726–729, 2021.
  • [19] B. Levy, Principles of Signal Detection and Parameter Estimation.   New York, USA: Springer, 2008.
  • [20] L. Kleinrock, Queueing systems: Volume 1-Theory, 5th ed.   John Wiley & Sons, 2018.
  • [21] E. Lehmann and J. Romano, Testing Statistical Hypotheses, 3rd ed.   Springer, 2005.
  • [22] H. Schaefer, Banach Lattices and Positive operators.   Berlin Heidelberg: Springer-Verlag, 1975.