This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On the Optimality of the Kautz-Singleton Construction in Probabilistic Group Testing

Huseyin A. Inan1, Peter Kairouz1, Mary Wootters2, and Ayfer Ozgur1 1Department of Electrical Engineering, Stanford University. hinan1, kairouzp, aozgur@stanford.edu2Departments of Computer Science and Electrical Engineering, Stanford University. marykw@stanford.eduThis work appeared in part at Allerton 2018.
Abstract

We consider the probabilistic group testing problem where dd random defective items in a large population of NN items are identified with high probability by applying binary tests. It is known that Θ(dlogN)\Theta(d\log N) tests are necessary and sufficient to recover the defective set with vanishing probability of error when d=O(Nα)d=O(N^{\alpha}) for some α(0,1)\alpha\in(0,1). However, to the best of our knowledge, there is no explicit (deterministic) construction achieving Θ(dlogN)\Theta(d\log N) tests in general. In this work, we show that a famous construction introduced by Kautz and Singleton for the combinatorial group testing problem (which is known to be suboptimal for combinatorial group testing for moderate values of dd) achieves the order optimal Θ(dlogN)\Theta(d\log N) tests in the probabilistic group testing problem when d=Ω(log2N)d=\Omega(\log^{2}N). This provides a strongly explicit construction achieving the order optimal result in the probabilistic group testing setting for a wide range of values of dd. To prove the order-optimality of Kautz and Singleton’s construction in the probabilistic setting, we provide a novel analysis of the probability of a non-defective item being covered by a random defective set directly, rather than arguing from combinatorial properties of the underlying code, which has been the main approach in the literature. Furthermore, we use a recursive technique to convert this construction into one that can also be efficiently decoded with only a log-log factor increase in the number of tests.

I Introduction

The objective of group testing is to efficiently identify a small set of dd defective items in a large population of size NN by performing binary tests on groups of items, as opposed to testing the items individually. A positive test outcome indicates that the group contains at least one defective item. A negative test outcome indicates that all the items in the group are non-defective. When dd is much smaller than NN, the defectives can be identified with far fewer than NN tests.

The original group testing framework was developed in 1943 by Robert Dorfman [1]. Back then, group testing was devised to identify which WWII draftees were infected with syphilis, without having to test them individually. In Dorfman’s application, items represented draftees and tests represented actual blood tests. Over the years, group testing has found numerous applications in fields spanning biology [2], medicine [3], machine learning [4], data analysis [5], signal processing [6], and wireless multiple-access communications [7, 8, 9, 10].

I-A Non-adaptive probabilistic group testing

Group testing strategies can be adaptive, where the ithi^{th} test is a function of the outcomes of the i1i-1 previous tests, or non-adaptive, where all tests are designed in one shot. A non-adaptive group testing strategy can be represented by a t×Nt\times N binary matrix M{M}, where Mij=1M_{ij}=1 indicates that item jj participates in test ii. Group testing schemes can also be combinatorial [11, 12] or probabilistic [13, 14, 15, 16, 17, 18, 19, 20].

The goal of combinatorial group testing schemes is to recover any set of up to dd defective items with zero-error and require at least t=min{N,Ω(d2logdN)}t=\min\{N,\Omega(d^{2}\log_{d}N)\} tests [21, 22]. A strongly explicit construction111We will call a t×Nt\times N matrix strongly explicit if any column of the matrix can be constructed in time poly(t)\textnormal{poly}(t). A matrix will be called explicit if it can be constructed in time poly(t,N)\textnormal{poly}(t,N). that achieves t=O(d2logd2N)t=O(d^{2}\log_{d}^{2}N) was introduced by Kautz and Singleton in [23]. A more recent explicit construction achieving t=O(d2logN)t=O(d^{2}\log N) was introduced by Porat and Rothschild in [24]. We note that the Kautz-Singleton construction matches the best known lower bound Ω(d2logdN)\Omega(d^{2}\log_{d}N) in the regime where d=Θ(Nα)d=\Theta(N^{\alpha}) for some α(0,1)\alpha\in(0,1). However, for moderate values of dd (e.g., d=O(poly(logN))d=O(\textnormal{poly}(\log N))), the construction introduced by Porat and Rothschild achieving t=O(d2logN)t=O(d^{2}\log N) is more efficient and the Kautz-Singleton construction is suboptimal in this regime.

In contrast, probabilistic group testing schemes assume a random defective set of size dd, allow for an arbitrarily small probability of reconstruction error, and require only t=Θ(dlogN)t=\Theta(d\log N) tests when d=O(N1α)d=O(N^{1-\alpha}) for some α(0,1)\alpha\in(0,1) [15, 16, 17]. In this paper, we are interested in non-adaptive probabilistic group testing schemes.

I-B Our contributions

To best of our knowledge, all known probabilistic group testing strategies that achieve t=O(dlogN)t=O(d\log N) tests are randomized (i.e., MM is randomly constructed) [13, 14, 15, 16, 17, 18, 19, 20]. Recently, Mazumdar [25] presented explicit schemes (deterministic constructions of MM) for probabilistic group testing framework. This was done by studying the average and minimum Hamming distances of constant-weight codes (such as Algebraic-Geometric codes) and relating them to the properties of group testing strategies. However, the explicit schemes in [25] achieve t=Θ(dlog2N/logd)t=\Theta(d\log^{2}N/\log d), which is not order-optimal when dd is poly-logarithmic in NN. It is therefore of interest to find explicit, deterministic schemes that achieve t=O(dlogN)t=O(d\log N) tests.

This paper presents a strongly explicit scheme that achieves t=O(dlogN)t=O(d\log N) in the regime where d=Ω(log2N)d=\Omega(\log^{2}N). We show that Kautz and Singleton’s well-known scheme is order-optimal for probabilistic group testing. This is perhaps surprising because for moderate values of dd (e.g., d=O(poly(logN))d=O(\textnormal{poly}(\log N))), this scheme is known to be sub-optimal for combinatorial group testing. We prove this result for both the noiseless and noisy (where test outcomes can be flipped at random) settings of probabilistic group testing framework. We prove the order-optimality of Kautz and Singleton’s construction by analyzing the probability of a non-defective item being “covered” (c.f. Section II) by a random defective set directly, rather than arguing from combinatorial properties of the underlying code, which has been the main approach in the literature [23, 24, 25].

We say a group testing scheme, which consists of a group testing strategy (i.e., MM) and a decoding rule, achieves probability of error ϵ\epsilon and is efficiently decodable if the decoding rule can identify the defective set in poly(t)\textnormal{poly}(t)-time complexity with ϵ\epsilon probability of error. While we can achieve the decoding complexity of O(tN)O(tN) with the “cover decoder” (c.f. Section II)222Common constructions in group testing literature have density Θ(1/d)\Theta(1/d), therefore, the decoding complexity can be brought to O(tN/d)O(tN/d)., our goal is to bring the decoding complexity to poly(t)\textnormal{poly}(t). To this end, we use a recursive technique inspired by [26] to convert the Kautz-Singleton construction into a strongly explicit construction with t=O(dlogNloglogdN)t=O(d\log N\log\log_{d}N) tests and decoding complexity O(d3logNloglogdN)O(d^{3}\log N\log\log_{d}N). This provides an efficiently decodable scheme with only a log-log factor increase in the number of tests. Searching for order-optimal explicit or randomized constructions that are efficiently decodable remains an open problem.

I-C Outline

The remainder of this paper is organized as follows. In Section II, we present the system model and necessary prerequisites. The optimality of the Kautz-Singleton construction in the probabilistic group testing setting is formally presented in Section III. We propose an efficiently decodable group testing strategy in Section IV. We defer the proofs of the results to their corresponding sections in the appendix. We provide, in Section V, a brief survey of related results on group testing and a detailed comparison with Mazumdar’s recent work in [25]. Finally, we conclude our paper in Section VI with a few interesting open problems.

II System Model and Basic Definitions

For any t×Nt\times N matrix M{M}, we use Mi{M}_{i} to refer to its ii’th column and Mij{M}_{ij} to refer to its (i,j)(i,j)’th entry. The support of a column MiM_{i} is the set of coordinates where MiM_{i} has nonzero entries. For an integer m1m\geq 1, we denote the set {1,,m}\{1,\ldots,m\} by [m][m]. The Hamming weight of a column of MM will be simply referred to as the weight of the column.

We consider a model where there is a random defective set SS of size dd among the items [N][N]. We define 𝒮\mathcal{S} as the set of all possible defective sets, i.e., the set of (Nd)\binom{N}{d} subsets of [N][N] of cardinality dd and we let SS be uniformly distributed over 𝒮\mathcal{S}.333This assumption is not critical. Our results carry over to the setting where the defective items are sampled with replacement. The goal is to determine SS from the binary measurement vector YY of size tt taking the form

Y=(iSMi)v,\displaystyle Y=\left(\bigvee_{i\in S}M_{i}\right)\oplus v, (1)

where t×Nt\times N measurement matrix MM indicates which items are included in the test, i.e., Mij=1{M}_{ij}=1 if the item jj is participated in test ii, v{0,1}tv\in\{0,1\}^{t} is a noise term, and \oplus denotes modulo-2 addition. In words, the measurement vector YY is the Boolean OR combination of the columns of the measurement matrix MM corresponding to the defective items in a possible noisy fashion. We are interested in both the noiseless and noisy variants of the model in (1). In the noiseless case, we simply consider v=0v=0, i.e., Y=iSMiY=\bigvee_{i\in S}M_{i}. Note that the randomness in the measurement vector YY is only due to the random defective set in this case. On the other hand, in the noisy case we consider vBernoulli(p)v\sim\textnormal{Bernoulli}(p) for some fixed constant p(0,0.5)p\in(0,0.5), i.e., each measurement is independently flipped with probability pp.

Given MM and YY, a decoding procedure forms an estimate S^\hat{S} of SS. The performance measure we consider in this paper is the exact recovery where the average probability of error is given by

PePr(S^S),\displaystyle P_{e}\triangleq\Pr(\hat{S}\neq S),

and is taken over the realizations of SS and vv (in the noisy case). The goal is to minimize the total number of tests tt while achieving a vanishing probability of error, i.e., satisfying Pe0P_{e}\rightarrow 0.

II-A Disjunctiveness

We say that a column MiM_{i} is covered by a set of columns Mj1,,MjlM_{j_{1}},\ldots,M_{j_{l}} with j1,,jl[N]{j_{1}},\ldots,{j_{l}}\in[N] if the support of MiM_{i} is contained in the union of the supports of columns Mj1,,MjlM_{j_{1}},\ldots,M_{j_{l}}. A binary matrix MM is called dd-disjunct if any column of MM is not covered by any other dd columns. The dd-disjunctiveness property ensures that we can recover any defective set of size dd with zero error from the measurement vector YY in the noiseless case. This can be naively done using the cover decoder (also referred as the COMP decoder [15, 17]) which runs in O(tN)O(tN)-time. The cover decoder simply scans through the columns of MM, and returns the ones that are covered by the measurement vector YY. When MM is dd-disjunct, the cover decoder succeeds at identifying all the defective items without any error.

In this work, we are interested in the probabilistic group testing problem where the 0-error condition is relaxed into a vanishing probability of error. Therefore we can relax the dd-disjunctiveness property. Note that to achieve an arbitrary but fixed ϵ\epsilon average probability of error in the noiseless case, it is sufficient to ensure that at least (1ϵ)(1-\epsilon) fraction of all possible defective sets do not cover any other column. A binary matrix satisfying this relaxed form is called an almost disjunct matrix [25, 27, 28, 29] and with this condition one can achieve the desired ϵ\epsilon average probability of error by applying the cover decoder.

II-B Kautz-Singleton Construction

In their work [23], Kautz and Singleton provide a construction of disjunct matrices by converting a Reed-Solomon (RS) code [30] to a binary matrix. We begin with the definition of Reed-Solomon codes.

Definition 1.

Let 𝔽q\mathbb{F}_{q} be a finite field and α1,,αn\alpha_{1},\ldots,\alpha_{n} be distinct elements from 𝔽q\mathbb{F}_{q}. Let knqk\leq n\leq q. The Reed-Solomon code of dimension kk over 𝔽q\mathbb{F}_{q}, with evaluation points α1,,αn\alpha_{1},\ldots,\alpha_{n} is defined with the following encoding function. The encoding of a message m=(m0,,mk1)m=(m_{0},\ldots,m_{k-1}) is the evaluation of the corresponding k1k-1 degree polynomial fm(X)=i=0k1miXif_{m}(X)=\sum_{i=0}^{k-1}m_{i}X^{i} at all the αi\alpha_{i}’s:

RS(m)=(fm(α1),,fm(αn)).\displaystyle\textnormal{RS}(m)=(f_{m}(\alpha_{1}),\ldots,f_{m}(\alpha_{n})).

The Kautz-Singleton construction starts with a [n,k]q[n,k]_{q} RS code with n=qn=q and N=qkN=q^{k}. Each qq-ary symbol is then replaced by a unit weight binary vector of length qq, via “identity mapping” which takes a symbol i[q]i\in[q] and maps it to the vector in {0,1}q\{0,1\}^{q} that has a 1 in the ii’th position and zero everywhere else. Note that the resulting binary matrix will have t=nq=q2t=nq=q^{2} tests. An example illustrating the Kautz-Singleton construction is depicted in Figure 1. This construction achieves a dd-disjunct t×Nt\times N binary matrix with t=O(d2logd2N)t=O(d^{2}\log_{d}^{2}N) by choosing the parameter qq appropriately. While the choice n=qn=q is appropriate for the combinatorial group testing problem, we will shortly see that we need to set n=Θ(logN)n=\Theta(\log N) in order to achieve the order-optimal result in the probabilistic group testing problem.

Refer to caption
Figure 1: An example of the Kautz-Singleton construction with [3,1]3[3,1]_{3} Reed-Solomon code.

While this is a strongly explicit construction, it is suboptimal for combinatorial group testing in the regime d=O(poly(logN))d=O(\textnormal{poly}(\log N)): an explicit construction with smaller tt (achieving t=O(d2logN)t=O(d^{2}\log N)) is introduced by Porat and Rothschild in [24]. Interestingly, we will show in the next section that this same strongly explicit construction that is suboptimal for combintorial group testing in fact achieves the order-optimal t=Θ(dlogN)t=\Theta(d\log N) result in both the noiseless and noisy versions of probabilistic group testing.

III Optimality of the Kautz-Singleton construction

We begin with the noiseless case (v=0v=0 in (1)). The next theorem shows the optimality of the Kautz-Singleton construction with properly chosen parameters nn and qq.

Theorem 1.

Let δ>0\delta>0. Under the noiseless model introduced in Section II, the Kautz-Singleton construction with parameters q=c1dq=c_{1}d for any c14c_{1}\geq 4 and n=(1+δ)logNn=(1+\delta)\log N has average probability of error PeNΩ(logq)+NδP_{e}\leq N^{-\Omega(\log q)}+N^{-\delta} under the cover decoder in the regime d=Ω(log2N)d=\Omega(\log^{2}N).

The proof of the above theorem can be found in Appendix -A. We note that the Kautz-Singleton construction in Theorem 1 has t=nq=Θ(dlogN)t=nq=\Theta(d\log N) tests, therefore, achieving the order-optimal result in the probabilistic group testing problem in the noiseless case. It is further possible to extend this result to the noisy setting where we consider vBernoulli(p)v\sim\textnormal{Bernoulli}(p) for some fixed constant p(0,0.5)p\in(0,0.5), i.e., each measurement is independently flipped with probability pp. Our next theorem shows the optimality of the Kautz-Singleton construction in this case.

Theorem 2.

Let δ>0\delta>0. Under the noisy model introduced in Section II with some fixed noise parameter p(0,0.5)p\in(0,0.5), the Kautz-Singleton construction with parameters q=c1dq=c_{1}d for any c124c_{1}\geq 24 and n=c2(1+δ)logNn=c_{2}(1+\delta)\log N for any c2max{89(0.5p)2,40.57}c_{2}\geq\max\{\frac{8}{9(0.5-p)^{2}},40.57\} has average probability of error PeNΩ(logq)+3NδP_{e}\leq N^{-\Omega(\log q)}+3N^{-\delta} under the modified version of cover decoder in the regime d=Ω(log2N)d=\Omega(\log^{2}N).

The proof of the above theorem can be found in Appendix -B. Similar to the noiseless setting, the Kautz-Singleton construction provides a strongly explicit construction achieving optimal number of tests t=Θ(dlogN)t=\Theta(d\log N) in the noisy case.

Given that the Kautz-Singleton construction achieves a vanishing probability of error with t=Θ(dlogN)t=\Theta(d\log N) order-optimal number of tests, a natural question of interest is how large the constant is and how the performance of this construction compares to random designs for given finite values of dd and NN. To illustrate the empirical performance of the Kautz-Singleton construction in the noiseless case, we provide simulation results in Figure 2 and 3 for different choices of NN and dd and compare the results to random designs considered in the literature. We used the code in [31] (see [32] for the associated article) for the Kautz-Singleton construction. For comparison, we take two randomized constructions from the literature, namely the Bernoulli design (see [17]) and the near-constant column weight design studied in [18]. We use the cover decoder for decoding. The simulation results illustrate that the Kautz-Singleton construction achieves better success probability for the same number of tests, which suggests that the implied constant for the Kautz-Singleton construction may be better than those for these random designs; we note that similar empirical findings were observed in [32]. Since the Kautz-Singleton construction additionally has an explicit and simple structure, this construction may be a good choice for designing measurement matrices for probabilistic group testing in practice.

Refer to caption
Figure 2: Empirical performances of the Kautz-Singleton construction along with the random near-constant column weight [18] and Bernoulli designs [17] under the cover decoder for N=500N=500 items and d=10d=10 defectives. For the Kautz-Singleton construction, empirical performance was judged using 5000 random trials and the number of tests correspond to a range of (q,n)(q,n) pair selections. For the random matrices, empirical performance was judged from 100 trials each on 100 random matrices.
Refer to caption
Figure 3: Empirical performances of the Kautz-Singleton construction along with the random near-constant column weight [18] and Bernoulli designs [17] under the cover decoder for N=2000N=2000 items and d=100d=100 defectives. For the Kautz-Singleton construction, empirical performance was judged using 5000 random trials and the number of tests correspond to a range of (q,n)(q,n) pair selections. For the random matrices, empirical performance was judged from 100 trials each on 100 random matrices.

IV Decoding

While the cover decoder, which has a decoding complexity of O(tN)O(tN), might be reasonable for certain applications, there is a recent research effort towards low-complexity decoding schemes due to the emerging applications involving massive datasets [33, 34, 26, 35, 36]. The target is a decoding complexity of poly(t)\textnormal{poly}(t). This is an exponential improvement in the running time over the cover decoder for moderate values of dd. For the model we consider in this work (i.e., exact recovery of the defective set with vanishing probability of error), there is no known efficiently decodable scheme with optimal t=Θ(dlogN)t=\Theta(d\log N) tests to the best of our knowledge. The work [35] presented a randomized scheme which identifies all the defective items with high probability with O(dlogdlogN)O(d\log d\log N) tests and time complexity O(dlogdlogN)O(d\log d\log N). Another recent result, [36], introduced an algorithm which requires O(dlogdlogN)O(d\log d\log N) tests with O(d(log2d+logN))O(d(\log^{2}d+\log N)) decoding complexity. Note that the decoding complexity reduces to O(dlogN)O(d\log N) when d=O(poly(logN))d=O(\textnormal{poly}(\log N)) which is order-optimal (and sub-linear in the number of tests), although the number of tests is not. In both [35] and [36], the number of tests is away from the optimal number of tests by a factor of logd\log d.

We can convert the strongly explicit constructions in Theorem 1 and 2 into strongly explicit constructions that are also efficiently decodable by using a recursive technique introduced in [26] where the authors construct efficiently decodable error-tolerant list disjunct matrices. For the sake of completeness, we next discuss the main idea applied to our case.

The cover decoder goes through the columns of MM and decides whether the corresponding item is defective or not. This results in decoding complexity O(tN)O(tN). Assume we were given a superset SS^{\prime} such that SS^{\prime} is guaranteed to include the defective set SS, i.e. SSS\subseteq S^{\prime}, then the cover decoder could run in time O(t|S|)O(t\cdot|S^{\prime}|) over the columns corresponding to SS^{\prime}, which depending on the size of SS^{\prime} could result in significantly lower complexity. It turns out that we can construct this small set SS^{\prime} recursively.

Suppose that we have access to an efficiently decodable t1(d,N,ϵ/4,p)×Nt_{1}(d,\sqrt{N},\epsilon/4,p)\times\sqrt{N} matrix M(1)M^{(1)} which can be used to detect at most dd defectives among N\sqrt{N} items with probability of error Peϵ/4P_{e}\leq{\epsilon/4} when the noise parameter is pp by using t1(d,N,ϵ/4,p)t_{1}({d},\sqrt{N},{\epsilon/4},{p}) tests. We construct two t1(d,N,ϵ/4,p)×Nt_{1}(d,\sqrt{N},\epsilon/4,p)\times N matrices M(F)M^{(F)} and M(L)M^{(L)} using M(1)M^{(1)} as follows. For i[N]i\in[N], the ii’th column of M(F)M^{(F)} is equal to jj’th column of M(1)M^{(1)} if the first 12logN\frac{1}{2}\log N bits in the binary representation of ii are given by the binary representation of jj for j[N]j\in[\sqrt{N}]. Similarly, for i[N]i\in[N], the ii’th columns of M(L)M^{(L)} is equal to the jj’th column of M(1)M^{(1)} if the last 12logN\frac{1}{2}\log N bits in the binary representation of ii are given by the binary representation of jj for j[N]j\in[\sqrt{N}].

The final matrix matrix MM is constructed by vertically stacking M(F)M^{(F)}, M(L)M^{(L)} and a t2(d,N,ϵ/2,p)×Nt_{2}(d,N,\epsilon/2,p)\times N matrix M(2)M^{(2)} which is not necessarily efficiently decodable (e.g., the Kautz-Singleton construction). As before, t2(d,N,ϵ/2,p)t_{2}(d,N,\epsilon/2,p) is the number of tests for M(2)M^{(2)}, which we assume can be used to detect dd defectives among NN items with probability of error Peϵ/2P_{e}\leq{\epsilon/2} when the noise parameter is pp. The decoding works as follows. We obtain the measurement vectors Y(F)Y^{(F)}, Y(L)Y^{(L)}, and Y(2)Y^{(2)} given by Y(F)=iSMi(F)v(F)Y^{(F)}=\bigvee_{i\in S}M^{(F)}_{i}\oplus v^{(F)}, Y(L)=iSMi(L)v(L)Y^{(L)}=\bigvee_{i\in S}M^{(L)}_{i}\oplus v^{(L)}, and Y(2)=iSMi(2)v(2)Y^{(2)}=\bigvee_{i\in S}M^{(2)}_{i}\oplus v^{(2)} respectively where v(F)v^{(F)}, v(L)v^{(L)}, and v(2)v^{(2)} are the noise terms corrupting the corresponding measurements. We next apply the decoding algorithm for M(1)M^{(1)} to Y(F)Y^{(F)} and Y(L)Y^{(L)} to obtain the estimate sets S^(F)\hat{S}^{(F)} and S^(L)\hat{S}^{(L)} respectively. Note that the sets S^(F)\hat{S}^{(F)} and S^(L)\hat{S}^{(L)} can decode the first and last 12logN\frac{1}{2}\log N-bits of the defective items respectively with probability at least 1ϵ/21-\epsilon/2 by the union bound. Therefore, we can construct the set S=S^(F)×S^(L)S^{\prime}=\hat{S}^{(F)}\times\hat{S}^{(L)} where ×\times denotes the Cartesian product and obtain a super set that contains the defective set SS with error probability at most ϵ/2\epsilon/2. We further note that since |S^(F)|d|\hat{S}^{(F)}|\leq d and |S^(L)|d|\hat{S}^{(L)}|\leq d, we have |S|d2|S^{\prime}|\leq d^{2}. We finally apply the naive cover decoder to M(2)M^{(2)} by running it over the set SS^{\prime} to compute the final estimate S^\hat{S} which can be done in additional O(t2d2)O(t_{2}\cdot d^{2}) time. Note that by the union bound the probability of error is bounded by ϵ\epsilon. Figure 4 illustrates the main idea with the example of d=2d=2 and N=16N=16. We provide this decoding scheme in Algorithm 1 for the special case N=d2iN=d^{2^{i}} for some non-negative integer ii although the results hold in the general case and no extra assumption beyond d=Ω(log2N)d=\Omega(\log^{2}N) is needed. The next theorem is the result of applying this idea recursively.

Refer to caption
Figure 4: An illustration of the construction presented in Section IV for the case d=2d=2 and N=16N=16. The illustration depicts the main idea, and the full construction is achieved by applying this idea recursively.
Input: The measurement vector YY, the group testing matrix MM, the defective set size dd, the number of items NN
Output: The defective set estimate S^\hat{S}
1 if N=dN=d then
2  Return the defective set using YY (individual testing);
3 
4else
5  Compute M(1)M^{(1)} and M(2)M^{(2)} (as described in the text);
6  Compute Y(F)Y^{(F)} and Y(L)Y^{(L)} (as described in the text);
7  S^(F)=decode(Y(F),M(1),d,N)\hat{S}^{(F)}=decode(Y^{(F)},M^{(1)},d,\sqrt{N});
8  S^(L)=decode(Y(L),M(1),d,N)\hat{S}^{(L)}=decode(Y^{(L)},M^{(1)},d,\sqrt{N});
9  if |S^(F)|>d|\hat{S}^{(F)}|>d or |S^(L)|>d|\hat{S}^{(L)}|>d then
10     return {};
11    
12 Construct S=S^(F)×S^(L)S^{\prime}=\hat{S}^{(F)}\times\hat{S}^{(L)};
13  Apply the cover decoder to M(2)M^{(2)} over the set SS^{\prime} and compute S^\hat{S};
14  Return S^\hat{S};
15 
Algorithm 1 The decoding alg. decode(YY, MM, dd, NN)

Theorem 3.

Under the noiseless/noisy model introduced in Section II, there exists a strongly explicit construction and a decoding rule achieving an arbitrary but fixed ϵ\epsilon average probability of error with t=O(dlogNloglogdN)t=O(d\log N\log\log_{d}N) number of tests that can be decoded in time O(d3logNloglogdN)O(d^{3}\log N\log\log_{d}N) in the regime d=Ω(log2N)d=\Omega(\log^{2}N).

The proof of the above theorem can be found in Appendix -C. We note that with only loglogdN\log\log_{d}N extra factor in the number of tests, the decoding complexity can be brought to the desired O(poly(t))O(\textnormal{poly}(t)) complexity. We further note that the number of tests becomes order-optimal in the regime d=Θ(Nα)d=\Theta(N^{\alpha}) for some α(0,1)\alpha\in(0,1). In Table I we provide the results presented in this work along with the related results in the literature.

Reference Number of tests Decoding complexity Construction
[17, 37] t=Θ(dlogN)t=\Theta(d\log N) O(tN)O(tN) Randomized
[25] t=O(dlog2N/logd)t=O(d\log^{2}N/\log d) O(tN)O(tN) Strongly explicit
[35] t=O(dlogdlogN)t=O(d\log d\log N) O(dlogdlogN)O(d\log d\log N) Randomized
[36] t=O(dlogdlogN)t=O(d\log d\log N) O(d(log2d+logN))O(d(\log^{2}d+\log N)) Randomized
This work t=Θ(dlogN)t=\Theta(d\log N) O(tN)O(tN) Strongly explicit
This work t=O(dlogNloglogdN)t=O(d\log N\log\log_{d}N) O(d3logNloglogdN)O(d^{3}\log N\log\log_{d}N) Strongly explicit
TABLE I: Comparison of non-adaptive probabilistic group testing results. We note that the main focus in [17, 37] is the implied constant in t=Θ(dlogN)t=\Theta(d\log N).

V Related Work

The literature on the non-adaptive group testing framework includes both explicit and random test designs. We refer the reader to [12] for a survey. In combinatorial group testing, the famous construction introduced by Kautz and Singleton [23] achieves t=O(d2logd2N)t=O(d^{2}\log_{d}^{2}N) tests matching the best known lower bound min{N,Ω(d2logdN)}\min\{N,\Omega(d^{2}\log_{d}N)\} [21, 22] in the regime where d=Θ(Nα)d=\Theta(N^{\alpha}) for some α(0,1)\alpha\in(0,1). However, this strongly explicit construction is suboptimal in the regime where d=O(poly(logN))d=O(\textnormal{poly}(\log N)). An explicit construction achieving t=O(d2logN)t=O(d^{2}\log N) was introduced by Porat and Rothschild in [24]. While t=O(d2logN)t=O(d^{2}\log N) is the best known achievability result in combinatorial group testing framework, there is no strongly explicit construction matching it to the best of our knowledge. Regarding efficient decoding, recently Indyk, Ngo and Rudra [34] introduced a randomized construction with t=O(d2log(N))t=O(d^{2}\log(N)) tests that could be decoded in time poly(t)\textnormal{poly}(t). Furthermore, the construction in [34] can be derandomized in the regime d=O(logN/loglogN)d=O(\log N/\log\log N). Later, Ngo, Porat and Rudra [26] removed the constraint on dd and provided an explicit construction that can be decoded in time poly(t)\textnormal{poly}(t). The main idea of [34] was to consider list-disjunct matrices; a similar idea was considered by Cheraghchi in [33], which obtained explicit constructions of non-adaptive group testing schemes that handle noisy tests and return a list of defectives that may include false positives.

There are various schemes relaxing the zero-error criteria in the group testing problem. For instance, the model mentioned above, where the decoder always outputs a small super-set of the defective items, was studied in [33, 38, 39, 40]. These constructions have efficient (poly(t)\textnormal{poly}(t)-time) decoding algorithms, and so can be used alongside constructions without sublinear time decoding algorithms to speed up decoding. Another framework where the goal is to recover at least a (1ϵ)(1-\epsilon)-fraction (for any arbitrarily small ϵ>0\epsilon>0) of the defective set with high probability was studied in [35] where the authors provided a scheme with order-optimal O(dlogN)O(d\log N) tests and the computational complexity. There are also different versions of the group testing problem in which a test can have more than two outcomes [41, 42] or can be threshold based [43, 44, 45]. More recently, sparse group testing frameworks for both combinatorial and probabilistic settings were studied in [46, 47, 48].

When the defective set is assumed to be uniformly random, it is known that t=Θ(dlogN)t=\Theta(d\log N) is order-optimal for achieving the exact recovery of the defective set with vanishing probability of error (which is the model considered in this work) in the broad regime d=O(Nα)d=O(N^{\alpha}) for some α(0,1)\alpha\in(0,1) using random designs and information-theoretical tools [37, 16]. These results also include the noisy variants of the group testing problem. Efficient recovery algorithms with nearly optimal number of tests were introduced recently in [35] and [36]. Regarding deterministic constructions of almost disjunct matrices, recently Mazumdar [25] introduced an analysis connecting the group testing properties with the average Hamming distance between the columns of the measurement matrix and obtained (strongly) explicit constructions with t=O(dlog2N/logd)t=O(d\log^{2}N/\log d) tests. While this result is order-optimal in the regime where d=Θ(Nα)d=\Theta(N^{\alpha}) for some α(0,1)\alpha\in(0,1), it is suboptimal for moderate values of dd (e.g., d=O(poly(logN))d=O(\textnormal{poly}(\log N))). The performance of the Kautz-Singleton construction in the random model has been studied empirically [32], but we are not aware of any theoretical analysis of it beyond what follows immediately from the distance of Reed-Solomon codes. To the best of our knowledge there is no known explicit/strongly explicit construction achieving t=Θ(dlogN)t=\Theta(d\log N) tests in general for the noiseless/noisy version of the probabilistic group testing problem.

VI Conclusion

In this work, we showed that the Kautz-Singleton construction is order-optimal in the noiseless and noisy variants of the probabilistic group testing problem. To the best of our knowledge, this is the first (strongly) explicit construction achieving order-optimal number of tests in the probabilistic group testing setting for poly-logarithmic (in NN) values of dd. We provided a novel analysis departing from the classical approaches in the literature that use combinatorial properties of the underlying code. We instead directly explored the probability of a non-defective item being covered by a random defective set using the properties of Reed-Solomon codes in our analysis. Furthermore, by using a recursive technique, we converted the Kautz-Singleton construction into a construction that is also efficiently decodable with only a log-log factor increase in number of tests which provides interesting tradeoffs compared to the existing results in the literature.

There are a number of nontrivial extensions to our work. Firstly, it would be interesting to extend these results to the regime d=o(log2N)d=o(\log^{2}N). Another interesting line of work would be to find a deterministic/randomized construction achieving order-optimal t=Θ(dlogN)t=\Theta(d\log N) tests and is also efficiently decodable.

-A Proof of Theorem 1

Let NN be the number of items and dd be the size of the random defective set. We will employ the Kautz-Singleton construction which takes a [n,k]q[n,k]_{q} RS code and replaces each qq-ary symbol by a unit weight binary vector of length qq using identity mapping. This corresponds to mapping a symbol i[q]i\in[q] to the vector in {0,1}q\{0,1\}^{q} that has a 1 in the ii’th position and zero everywhere else (see Section II-B for the full description). Note that the resulting t×Nt\times N binary matrix MM has t=nqt=nq tests. We shall later see that the choice q=4dq=4d and n=Θ(logN)n=\Theta(\log N) is appropriate, therefore, leading to t=Θ(dlogN)t=\Theta(d\log N) tests.

We note that for any defective set the cover decoder provides an exact recovery given that none of the non-defective items are covered by the defective set. Recall that a column MiM_{i} is covered by a set of columns Mj1,,MjlM_{j_{1}},\ldots,M_{j_{l}} with j1,,jl[N]{j_{1}},\ldots,{j_{l}}\in[N] if the support of MiM_{i} is contained in the union of the supports of columns Mj1,,MjlM_{j_{1}},\ldots,M_{j_{l}}. Note that in the noiseless case the measurement vector YY is given by the Boolean OR of the columns corresponding to the defective items. Therefore, the measurement vector YY covers all defective items, and the cover decoder can achieve exact recovery if none of the non-defective items are covered by the measurement vector YY (or equivalently the defective set).

For s[N]s\subseteq[N], we define 𝒜s\mathcal{A}^{s} as the event that there exists a non-defective column of MM that is covered by the defective set ss. Define 𝒜is\mathcal{A}_{i}^{s} as the event that the non-defective column MiM_{i} (isi\notin s) is covered by the defective set ss. We can bound the probability of error as follows:

Pe\displaystyle P_{e} s[N],|s|=d1(𝒜s)Pr(S=s)\displaystyle\leq\sum\limits_{s\subseteq[N],|s|=d}1(\mathcal{A}^{s})\Pr(S=s)
1(Nd)s[N],|s|=di[N]\s1(𝒜is)\displaystyle\leq\dfrac{1}{\binom{N}{d}}\sum\limits_{s\subseteq[N],|s|=d}\ \sum\limits_{i\in[N]\backslash s}1(\mathcal{A}_{i}^{s})
=1(Nd)i[N]s[N]/{i},|s|=d1(𝒜is)\displaystyle=\dfrac{1}{\binom{N}{d}}\sum\limits_{i\in[N]}\ \sum\limits_{s\subseteq[N]/\{i\},|s|=d}1(\mathcal{A}_{i}^{s})
=(N1d)(Nd)i[N]1(N1d)s[N]/{i},|s|=d1(𝒜is)\displaystyle=\dfrac{\binom{N-1}{d}}{\binom{N}{d}}\sum\limits_{i\in[N]}\ \dfrac{1}{\binom{N-1}{d}}\sum\limits_{s\subseteq[N]/\{i\},|s|=d}1(\mathcal{A}_{i}^{s})
=NdNi[N]Pr(𝒜iS[N]/{i})\displaystyle=\dfrac{N-d}{N}\sum\limits_{i\in[N]}\Pr\left(\mathcal{A}_{i}^{S_{[N]/\{i\}}}\right) (2)

where in the last equation S[N]/{i}S_{[N]/\{i\}} is uniformly distributed on the sets of size dd among the items in [N]/{i}[N]/\{i\} and 1()1(\cdot) denotes the indicator function of an event.

Fix any nn distinct elements α1,α2,,αn\alpha_{1},\alpha_{2},\ldots,\alpha_{n} from 𝔽q\mathbb{F}_{q}. We denote Ψ{α1,α2,,αn}\Psi\triangleq\{\alpha_{1},\alpha_{2},\ldots,\alpha_{n}\}. We note that due to the structure of mapping to the binary vectors in the Kautz-Singleton construction, a column MiM_{i} is covered by the random defective set SS if and only if the corresponding symbols of MiM_{i} are contained in the union of symbols of SS in the RS code for all rows in [n][n]. Recall that there is a k1k-1 degree polynomial fm(X)=i=0k1miXif_{m}(X)=\sum_{i=0}^{k-1}m_{i}X^{i} corresponding to each column in the RS code and the corresponding symbols in the column are the evaluation of fm(X)f_{m}(X) at α1,α2,,αn\alpha_{1},\alpha_{2},\ldots,\alpha_{n}. Denoting fmi(X)f_{m_{i}}(X) as the polynomial corresponding to the column MiM_{i}, we have

Pr(𝒜iS[N]/{i})\displaystyle\Pr\left(\mathcal{A}_{i}^{S_{[N]/\{i\}}}\right) =Pr(fmi(α){fmj(α):jS[N]/{i}}αΨ)\displaystyle=\Pr\left(f_{m_{i}}(\alpha)\in\left\{f_{m_{j}}(\alpha):j\in S_{[N]/\{i\}}\right\}\ \forall\ \alpha\in\Psi\right)
=Pr(0{fmj(α)fmi(α):jS[N]/{i}}αΨ).\displaystyle=\Pr\left(0\in\left\{f_{m_{j}}(\alpha)-f_{m_{i}}(\alpha):j\in S_{[N]/\{i\}}\right\}\ \forall\ \alpha\in\Psi\right).

We note that the columns of the RS code contain all possible (at most) k1k-1 degree polynomials, therefore, the set {fmj(α)fmi(α):j[N]/{i}}\left\{f_{m_{j}}(\alpha)-f_{m_{i}}(\alpha):j\in{[N]/\{i\}}\right\} is sweeping through all possible (at most) k1k-1 degree polynomials except the zero polynomial. Therefore, the randomness of S[N]/{i}S_{[N]/\{i\}} that generates the random set {fmj(α)fmi(α):jS[N]/{i}}\left\{f_{m_{j}}(\alpha)-f_{m_{i}}(\alpha):j\in S_{[N]/\{i\}}\right\} can be translated to the random set of polynomials {fmj(X):jS}\{f_{m_{j}}(X):j\in S^{\prime}\} that is generated by picking dd nonzero polynomials of degree (at most) k1k-1 without replacement. This gives

Pr(0{fmj(α)fmi(α):jS[N]/{i}}αΨ)=Pr(0{fmj(α):jS}αΨ).\displaystyle\Pr\left(0\in\left\{f_{m_{j}}(\alpha)-f_{m_{i}}(\alpha):j\in S_{[N]/\{i\}}\right\}\ \forall\ \alpha\in\Psi\right)=\Pr\left(0\in\left\{f_{m_{j}}(\alpha):j\in S^{\prime}\right\}\ \forall\ \alpha\in\Psi\right).

We define the random polynomial h(X)jSfmj(X)h(X)\triangleq\prod\limits_{j\in S^{\prime}}f_{m_{j}}(X). Note that

0{fmj(α):jS}αΨh(α)=0αΨ.\displaystyle 0\in\{f_{m_{j}}(\alpha):j\in S^{\prime}\}\ \forall\ \alpha\in\Psi\ \Leftrightarrow\ h(\alpha)=0\ \forall\ \alpha\in\Psi.

We next bound the number of roots of the polynomial h(X)h(X). We will use the following result from [49].

Lemma 1 ([49, Lemma 3.9]).

Let Rq(l,k)R_{q}(l,k) denote the set of nonzero polynomials over 𝔽q\mathbb{F}_{q} of degree at most kk that have exactly ll distinct roots in 𝔽q\mathbb{F}_{q}. For all powers qq and integers l,k,l,k,

|Rq(l,k)|qk+11l!.\displaystyle|R_{q}(l,k)|\leq q^{k+1}\cdot\dfrac{1}{l!}.

Let rr denote the number of roots of a random nonzero polynomial of degree at most k1k-1. One can observe that 𝔼[r]1\mathbb{E}[r]\leq 1 by noting that there is exactly one value of m0m_{0} that makes fm(X)=0f_{m}(X)=0 for any fixed XX and m1,,mk1m_{1},\ldots,m_{k-1} and the inequality is due to excluding the zero polynomial. Furthermore, using Lemma 1, we get

𝔼[r2]\displaystyle\mathbb{E}[r^{2}] i=1k1i2i!\displaystyle\leq\sum\limits_{i=1}^{k-1}\dfrac{i^{2}}{i!}
=i=1k1i(i1)!\displaystyle=\sum\limits_{i=1}^{k-1}\dfrac{i}{(i-1)!}
=i=1k1i1(i1)!+i=1k11(i1)!\displaystyle=\sum\limits_{i=1}^{k-1}\dfrac{i-1}{(i-1)!}+\sum\limits_{i=1}^{k-1}\dfrac{1}{(i-1)!}
<2e\displaystyle<2e

where the first inequality is due to Pr(r=i)=|Rq(i,k1)|/qk1/i!\Pr(r=i)=|R_{q}(i,k-1)|/q^{k}\leq 1/i! from Lemma 1. Hence we can bound 𝔼[r2]<6\mathbb{E}[r^{2}]<6. We denote rir_{i} as the number of roots of the polynomial fmi(X)f_{m_{i}}(X) and rhr_{h} as the number of roots of the polynomial h(X)h(X). Note that rhjSrjr_{h}\leq\sum_{j\in S^{\prime}}r_{j}. We will use the following Bernstein concentration bound for sampling without replacement [50]:

Proposition 1 ([50, Proposition 1.4]).

Let 𝒳={x1,,xN}\mathcal{X}=\{x_{1},\ldots,x_{N}\} be a finite population of NN points and X1,,XnX_{1},\ldots,X_{n} be a random sample drawn without replacement from 𝒳\mathcal{X}. Let a=min1iNxia=\min\limits_{1\leq i\leq N}x_{i} and b=max1iNxib=\max\limits_{1\leq i\leq N}x_{i}. Then for all ϵ>0\epsilon>0,

Pr(1ni=1nXiμ>ϵ)exp(nϵ22σ2+(2/3)(ba)ϵ)\Pr\left(\dfrac{1}{n}\sum\limits_{i=1}^{n}X_{i}-\mu>\epsilon\right)\leq\exp\left(-\dfrac{n\epsilon^{2}}{2\sigma^{2}+(2/3)(b-a)\epsilon}\right)

where μ=1Ni=1Nxi\mu=\frac{1}{N}\sum_{i=1}^{N}x_{i} is the mean of 𝒳\mathcal{X} and σ2=1Ni=1N(xiμ)2\sigma^{2}=\frac{1}{N}\sum_{i=1}^{N}(x_{i}-\mu)^{2} is the variance of 𝒳\mathcal{X}.

We apply the inequality above to jSrj\sum_{j\in S^{\prime}}r_{j} and obtain

Pr(jSrj>2d)\displaystyle\Pr\left(\sum\limits_{j\in S^{\prime}}r_{j}>2d\right) =Pr(1djSrj>2)\displaystyle=\Pr\left(\dfrac{1}{d}\sum\limits_{j\in S^{\prime}}r_{j}>2\right)
Pr(1djS(rj𝔼[rj])>1)\displaystyle\leq\Pr\left(\dfrac{1}{d}\sum\limits_{j\in S^{\prime}}(r_{j}-\mathbb{E}[r_{j}])>1\right)
exp(d12+k(2/3))\displaystyle\leq\exp\left(-\dfrac{d}{12+k(2/3)}\right)
exp(d16k).\displaystyle\leq\exp\left(-\dfrac{d}{16k}\right).

We have k=logN/logqk=\log N/\log q, hence, under the regime d=Ω(log2N)d=\Omega(\log^{2}N), the last quantity is bounded by NclogqN^{-c\log q} for some constant c>0c>0. Hence the number of roots of the polynomial h(X)h(X) is bounded by 2d2d with high probability.

Given the condition that the number of roots of the polynomial h(X)h(X) is bounded by 2d2d and the random set of polynomials {fmj(X):jS}\{f_{m_{j}}(X):j\in S^{\prime}\} is picked from the nonzero polynomials of degree at most k1k-1 without replacement, due to the symmetry in the position of the roots of the randomly selected polynomials, we claim that the probability of satisfying h(α)=0h(\alpha)=0 for all αΨ\alpha\in\Psi is bounded by the probability of covering nn elements from a field of size qq by picking 2d2d elements randomly without replacement. We next prove this claim. We define the set R(h){α𝔽q:h(α)=0}R(h)\coloneqq\{\alpha\in\mathbb{F}_{q}:h(\alpha)=0\} and we emphasize that this is not a multiset, i.e., the repeated roots appear as a single element. We begin with the following observation.

Claim 1.

Let l>0l>0, and condition on the event that |R(h)|=l|R(h)|=l. Then R(h)R(h) is uniformly distributed among all sets Λ𝔽q\Lambda\subseteq\mathbb{F}_{q} of size ll.

Proof.

For f𝔽q[X]f\in\mathbb{F}_{q}[X], we can write

f(X)=gf(X)γiR(f)(Xγi)ci,\displaystyle f(X)=g_{f}(X)\cdot\prod\limits_{\gamma_{i}\in R(f)}(X-\gamma_{i})^{c_{i}},

where cic_{i} is the corresponding multiplicity of the root γi\gamma_{i} and gf𝔽q[X]g_{f}\in\mathbb{F}_{q}[X] does not have any linear factor. We note that this decomposition is unique. For Λ𝔽q\Lambda\subseteq\mathbb{F}_{q} of size ll, let

HΛ{{f1(X),,fd(X)}:R(ifi(X))=Λ}.\displaystyle H_{\Lambda}\coloneqq\left\{\{f_{1}(X),\ldots,f_{d}(X)\}:R\left(\prod_{i}f_{i}(X)\right)=\Lambda\right\}.

Let Λ𝔽q\Lambda^{\prime}\subseteq\mathbb{F}_{q} such that |Λ|=l|\Lambda^{\prime}|=l and ΛΛ\Lambda^{\prime}\neq\Lambda. Then |HΛ|=|HΛ||H_{\Lambda}|=|H_{\Lambda^{\prime}}|. Indeed, let φ:𝔽q𝔽q\varphi:\mathbb{F}_{q}\rightarrow\mathbb{F}_{q} be a bijection such that φ(Λ)=Λ\varphi(\Lambda)=\Lambda^{\prime}. Then Φ:HΛHΛ\Phi:H_{\Lambda}\rightarrow H_{\Lambda^{\prime}} given by

Φ(f)=gf(X)γiR(f)(Xφ(γi))ci,\displaystyle\Phi(f)=g_{f}(X)\cdot\prod\limits_{\gamma_{i}\in R(f)}(X-\varphi(\gamma_{i}))^{c_{i}},

and Φ({f1,,fd}){Φ(f1),,Φ(fd)}\Phi(\{f_{1},\ldots,f_{d}\})\coloneqq\{\Phi(f_{1}),\ldots,\Phi(f_{d})\} is a bijection.

We further note that R(h)=Λ|R(h)|=lR(h)=\Lambda\Rightarrow|R(h)|=l, so

Pr{R(h)=Λ|R(h)|=l}\displaystyle\Pr\{R(h)=\Lambda\>\big{|}\>|R(h)|=l\} =Pr{R(h)=Λ}Pr{|R(h)|=l}\displaystyle=\dfrac{\Pr\{R(h)=\Lambda\}}{\Pr\{|R(h)|=l\}}
=Pr{{f1,,fd}HΛ}Pr{|R(h)|=l}\displaystyle=\dfrac{\Pr\{\{f_{1},\ldots,f_{d}\}\in H_{\Lambda}\}}{\Pr\{|R(h)|=l\}}
=(i)Pr{{f1,,fd}HΛ}Pr{|R(h)|=l}\displaystyle\stackrel{{\scriptstyle(i)}}{{=}}\dfrac{\Pr\{\{f_{1},\ldots,f_{d}\}\in H_{\Lambda^{\prime}}\}}{\Pr\{|R(h)|=l\}}
=Pr{R(h)=Λ|R(h)|=l},\displaystyle=\Pr\{R(h)=\Lambda^{\prime}\>\big{|}\>|R(h)|=l\},

where (i)(i) is due to |HΛ|=|HΛ||H_{\Lambda}|=|H_{\Lambda^{\prime}}| and we pick f1,,fdf_{1},\ldots,f_{d} uniformly without replacement. ∎

Based on this, if we ensure n2dn\leq 2d, then it follows that

Pr{R(h)Ψ|R(h)|2d}\displaystyle\Pr\{R(h)\supseteq\Psi\>\big{|}\>|R(h)|\leq 2d\}
=l2dPr{R(h)Ψ|R(h)|=l}Pr{|R(h)|=l|R(h)|2d}\displaystyle=\sum\limits_{l\leq 2d}\Pr\{R(h)\supseteq\Psi\>\big{|}\>|R(h)|=l\}\Pr\{|R(h)|=l\>\big{|}\>|R(h)|\leq 2d\}
maxnl2dPr{R(h)Ψ|R(h)|=l}\displaystyle\leq\max\limits_{n\leq l\leq 2d}\Pr\{R(h)\supseteq\Psi\>\big{|}\>|R(h)|=l\}
=maxnl2d(qnln)(ql).\displaystyle=\max\limits_{n\leq l\leq 2d}\dfrac{\binom{q-n}{l-n}}{\binom{q}{l}}.

Let us fix q=4dq=4d. We then have

Pr{R(h)Ψ|R(h)|2d}\displaystyle\Pr\{R(h)\supseteq\Psi\>\big{|}\>|R(h)|\leq 2d\} (4dn2dn)(4d2d)\displaystyle\leq\dfrac{\binom{4d-n}{2d-n}}{\binom{4d}{2d}}
=(4dn)!(2dn)!(2d)!(2d)!(2d)!(4d)!\displaystyle=\dfrac{(4d-n)!}{(2d-n)!(2d)!}\dfrac{(2d)!(2d)!}{(4d)!}
=2d(2dn+1)4d(4dn+1)\displaystyle=\dfrac{2d\ldots(2d-n+1)}{4d\ldots(4d-n+1)}
(12)n.\displaystyle\leq\left(\dfrac{1}{2}\right)^{n}.

Therefore, Pr(𝒜iS)\Pr(\mathcal{A}_{i}^{S}) is bounded by

Pr(𝒜iS)\displaystyle\Pr(\mathcal{A}_{i}^{S}) Pr{R(h)Ψ|R(h)|2d}+Pr{|R(h)|>2d}\displaystyle\leq\Pr\{R(h)\supseteq\Psi\>\big{|}\>|R(h)|\leq 2d\}+\Pr\{|R(h)|>2d\}
(12)n+Nclogq.\displaystyle\leq\left(\dfrac{1}{2}\right)^{n}+N^{-c\log q}.

Applying the summation over all i[N]i\in[N] in (2), we obtain PeN1clogq+N2nP_{e}\leq N^{1-c\log q}+N2^{-n}. Therefore, under the regime d=Ω(log2N)d=\Omega(\log^{2}N), the average probability of error can be bounded as PeNΩ(logq)+NδP_{e}\leq N^{-\Omega(\log q)}+N^{-\delta} by choosing n=(1+δ)logNn=(1+\delta)\log N. The condition n2dn\leq 2d required in the proof is also satisfied under this regime. Note that the resulting t×Nt\times N binary matrix MM has t=nq=Θ(dlogN)t=nq=\Theta(d\log N) tests.

-B Proof of Theorem 2

We begin with describing the decoding rule. Since we are considering the noisy model, we will slightly modify the cover decoder employed in the noiseless case. For any defective item with codeword weight ww, in the noiseless outcome the tests in which this item participated will be all positive. On the other hand, when the noise is added, wpwp of these tests will flip in expectation. Based on this observation (see No-CoMa in [37] for a more detailed discussion), we consider the following decoding rule. For any item i[N]i\in[N], we first denote wiw_{i} as the weight of the corresponding column MiM_{i} and w^i\hat{w}_{i} as the number of rows k[t]k\in[t] where both Mk,i=1M_{k,i}=1 and Yk=1Y_{k}=1. If w^iwi(1p(1+τ))\hat{w}_{i}\geq w_{i}(1-p(1+\tau)), then the iith item is declared as defective, else it is declared to be non-defective.

Under the aforementioned decoding rule, an error event happens either when w^i<wi(1p(1+τ))\hat{w}_{i}<w_{i}(1-p(1+\tau)) for a defective item ii or w^iwi(1p(1+τ))\hat{w}_{i}\geq w_{i}(1-p(1+\tau)) for a non-defective item ii. Using the union bound, we can bound the probability of error as follows:

Pe\displaystyle P_{e} 1(Nd)s[N],|s|=d[i[N]\sPr{w^iwi(1p(1+τ))}+isPr{w^i<wi(1p(1+τ))}]\displaystyle\leq\dfrac{1}{\binom{N}{d}}\sum\limits_{s\subseteq[N],|s|=d}\bigg{[}\ \sum\limits_{i\in[N]\backslash s}\Pr\{\textnormal{$\hat{w}_{i}\geq w_{i}(1-p(1+\tau))$}\}+\sum\limits_{i\in s}\Pr\{\textnormal{$\hat{w}_{i}<w_{i}(1-p(1+\tau))$}\}\bigg{]}
=1(Nd)i[N]s[N]/{i},|s|=dPr{w^iwi(1p(1+τ))}+1(Nd)s[N],|s|=disPr{w^i<wi(1p(1+τ))}\displaystyle=\dfrac{1}{\binom{N}{d}}\sum\limits_{i\in[N]}\ \sum\limits_{s\subseteq[N]/\{i\},|s|=d}\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))\}+\dfrac{1}{\binom{N}{d}}\sum\limits_{s\subseteq[N],|s|=d}\ \sum\limits_{i\in s}\Pr\{\hat{w}_{i}<w_{i}(1-p(1+\tau))\}
=(N1d)(Nd)(i[N]1(N1d)s[N]/{i},|s|=dPr{w^iwi(1p(1+τ))})\displaystyle=\dfrac{\binom{N-1}{d}}{\binom{N}{d}}\bigg{(}\sum\limits_{i\in[N]}\ \dfrac{1}{\binom{N-1}{d}}\sum\limits_{s\subseteq[N]/\{i\},|s|=d}\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))\}\bigg{)}
+1(Nd)s[N],|s|=disPr{w^i<wi(1p(1+τ))}\displaystyle\hskip 231.26378pt+\dfrac{1}{\binom{N}{d}}\sum\limits_{s\subseteq[N],|s|=d}\ \sum\limits_{i\in s}\Pr\{\hat{w}_{i}<w_{i}(1-p(1+\tau))\}
=NdNi[N]Pr{w^iwi(1p(1+τ))}+1(Nd)s[N],|s|=disPr{w^i<wi(1p(1+τ))}\displaystyle=\dfrac{N-d}{N}\sum\limits_{i\in[N]}\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))\}+\dfrac{1}{\binom{N}{d}}\sum\limits_{s\subseteq[N],|s|=d}\ \sum\limits_{i\in s}\Pr\{\hat{w}_{i}<w_{i}(1-p(1+\tau))\} (3)
P1+P2,\displaystyle\eqqcolon P_{1}+P_{2},

where we denote the first term of (3) as P1P_{1} and the second one as P2P_{2} in the last equation. We point out that in the first term of (3) the randomness is both due to the noise and the defective set that is uniformly distributed among the items in [N]/{i}[N]/\{i\} whereas in the second term the randomness is due to the noise.

We will employ the Kautz-Singleton construction which takes a [n,k]q[n,k]_{q} RS code and replaces each qq-ary symbol by unit weight binary vectors of length qq using identity mapping. This corresponds to mapping a symbol i[q]i\in[q] to the vector in {0,1}q\{0,1\}^{q} that has a 1 in the ii’th position and zero everywhere else (see Section II-B for the full description). Note that the resulting t×Nt\times N binary matrix MM has t=nqt=nq tests. We shall later see that the choice q=24dq=24d and n=Θ(logN)n=\Theta(\log N) is appropriate, therefore, leading to t=Θ(dlogN)t=\Theta(d\log N) tests. Fix any nn distinct elements α1,α2,,αn\alpha_{1},\alpha_{2},\ldots,\alpha_{n} from 𝔽q\mathbb{F}_{q}. We denote Ψ{α1,α2,,αn}\Psi\triangleq\{\alpha_{1},\alpha_{2},\ldots,\alpha_{n}\}.

We begin with P2P_{2}. Fix any defective set ss in [N][N] with size dd and fix an arbitrary element ii of this set. We first note that wi=nw_{i}=n due to the structure of the Kautz-Singleton construction. We further note that before the addition of noise the noiseless outcome will have positive entries corresponding to the ones where Mk,i=1M_{k,i}=1. Therefore Pr{w^i<wi(1p(1+τ))}\Pr\{\hat{w}_{i}<w_{i}(1-p(1+\tau))\} only depends on the number of bit flips due to the noise. Using Hoeffding’s inequality, we have

Pr{w^i<wi(1p(1+τ))}e2np2τ2.\displaystyle\Pr\{\hat{w}_{i}<w_{i}(1-p(1+\tau))\}\leq e^{-2np^{2}\tau^{2}}.

Summing over the dd defective items isi\in s, we get P2de2np2τ2P_{2}\leq de^{-2np^{2}\tau^{2}}.

We continue with P1P_{1}. We fix an item i[N]i\in[N] and note that wi=nw_{i}=n. We similarly define the random polynomial h(X)jSfmj(X)h(X)\triangleq\prod\limits_{j\in S}f_{m_{j}}(X). Let 𝒜\mathcal{A} be the event of h(X)h(X) having at most 2d2d number of roots. We then have

Pr{w^iwi(1p(1+τ))}\displaystyle\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))\} =Pr{w^iwi(1p(1+τ))|𝒜}Pr{𝒜}+Pr{w^iwi(1p(1+τ))|𝒜c}Pr{𝒜c}\displaystyle=\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A}\}\Pr\{\mathcal{A}\}+\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A}^{c}\}\Pr\{\mathcal{A}^{c}\}
Pr{w^iwi(1p(1+τ))|𝒜}+Pr{𝒜c}.\displaystyle\leq\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A}\}+\Pr\{\mathcal{A}^{c}\}. (4)

Following similar steps as in the proof of Theorem 1 we obtain Pr{𝒜c}Nclogq\Pr\{\mathcal{A}^{c}\}\leq N^{-c\log q} for some constant c>0c>0 in the regime d=Ω(log2N)d=\Omega(\log^{2}N).

We next bound the first term in (4). We choose q=24dq=24d and define the random set Υ={αΨ:fmi(α){fmj(α):jS}}\Upsilon=\{\alpha\in\Psi:f_{m_{i}}(\alpha)\in\{f_{m_{j}}(\alpha):j\in S\}\}. We then have

Pr{w^iwi(1p(1+τ))|𝒜}\displaystyle\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A}\} =Pr{w^iwi(1p(1+τ))|𝒜,|Υ|n/4}Pr{|Υ|n/4|𝒜}\displaystyle=\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A},|\Upsilon|\leq n/4\}\Pr\{|\Upsilon|\leq n/4|\mathcal{A}\}
+Pr{w^iwi(1p(1+τ))|𝒜,|Υ|>n/4}Pr{|Υ|>n/4|𝒜}\displaystyle\hskip 72.26999pt+\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A},|\Upsilon|>n/4\}\Pr\{|\Upsilon|>n/4|\mathcal{A}\}
Pr{w^iwi(1p(1+τ))|𝒜,|Υ|n/4}+Pr{|Υ|>n/4|𝒜}.\displaystyle\leq\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A},|\Upsilon|\leq n/4\}+\Pr\{|\Upsilon|>n/4|\mathcal{A}\}.

Let us first bound the second term Pr{|Υ|>n/4|𝒜}\Pr\{|\Upsilon|>n/4|\mathcal{A}\}. We note that

|Υ|\displaystyle|\Upsilon| =|{αΨ:fmi(α){fmj(α):jS}}|\displaystyle=|\{\alpha\in\Psi:f_{m_{i}}(\alpha)\in\{f_{m_{j}}(\alpha):j\in S\}\}|
=|{αΨ:0{fmj(α)fmi(α):jS}}|\displaystyle=|\{\alpha\in\Psi:0\in\{f_{m_{j}}(\alpha)-f_{m_{i}}(\alpha):j\in S\}\}|
=|{αΨ:0{fmj(α):jS}}|\displaystyle=|\{\alpha\in\Psi:0\in\{f_{m_{j}}(\alpha):j\in S^{\prime}\}\}|

where in the last equality the random set of polynomials {fmj(X):jS}\{f_{m_{j}}(X):j\in S^{\prime}\} is generated by picking dd nonzero polynomials of degree at most k1k-1 without replacement. This holds since iSi\notin S and the columns of the RS code contain all possible (at most) k1k-1 degree polynomials, therefore, the randomness of {fmj(α)fmi(α):jS}\{f_{m_{j}}(\alpha)-f_{m_{i}}(\alpha):j\in S\} can be translated to the random set of polynomials {fmj(X):jS}\{f_{m_{j}}(X):j\in S^{\prime}\} that is generated by picking dd nonzero polynomials of degree (at most) k1k-1 without replacement. Following similar steps of the proof of Theorem 1 we can bound Pr{|Υ|>n/4|𝒜}\Pr\{|\Upsilon|>n/4|\mathcal{A}\} by considering the probability of having at least n/4n/4 symbols from Ψ\Psi when we pick 2d2d symbols from [q][q] uniformly at random without replacement. Hence, if we ensure n8dn\leq 8d, then we have

Pr{|Υ|>n/4|𝒜}\displaystyle\Pr\{|\Upsilon|>n/4|\mathcal{A}\} (nn/4)(qn/42dn/4)(q2d)\displaystyle\leq\dfrac{\binom{n}{n/4}\binom{q-n/4}{2d-n/4}}{\binom{q}{2d}}
=(nn/4)(24dn/42dn/4)(24d2d)\displaystyle=\dfrac{\binom{n}{n/4}\binom{24d-n/4}{2d-n/4}}{\binom{24d}{2d}}
(4e)n/4(24dn/4)!(2dn/4)!(22d)!(2d)!(22d)!(24d)!\displaystyle\leq(4e)^{n/4}\dfrac{(24d-n/4)!}{(2d-n/4)!(22d)!}\dfrac{(2d)!(22d)!}{(24d)!}
=(4e)n/42d(2d1)(2dn/4+1)24d(24d1)(24dn/4+1)\displaystyle=(4e)^{n/4}\dfrac{2d(2d-1)\ldots(2d-n/4+1)}{24d(24d-1)\ldots(24d-n/4+1)}
(4e12)n/4\displaystyle\leq\left(\dfrac{4e}{12}\right)^{n/4}

where we use (nk)(en/k)k\binom{n}{k}\leq(en/k)^{k} in the second inequality.

We continue with Pr{w^iwi(1p(1+τ))|𝒜,|Υ|n/4}\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A},|\Upsilon|\leq n/4\}. Note that wi=nw_{i}=n. We further note that

𝔼[w^i]=𝔼[𝔼[w^i|Υ]]=𝔼[|Υ|](1p)+(n𝔼[|Υ|])p.\displaystyle\mathbb{E}[\hat{w}_{i}]=\mathbb{E}[\mathbb{E}[\hat{w}_{i}|\Upsilon]]=\mathbb{E}[|\Upsilon|](1-p)+(n-\mathbb{E}[|\Upsilon|])p.

Since p(0,0.5)p\in(0,0.5) we have 𝔼[w^i||Υ|n/4](n/4)(1p)+(3n/4)p=n/4+(n/2)p\mathbb{E}[\hat{w}_{i}\ |\ |\Upsilon|\leq n/4]\leq(n/4)(1-p)+(3n/4)p=n/4+(n/2)p. Using Hoeffding’s inequality, we have

Pr{w^iwi(1p(1+τ))|𝒜,|Υ|n/4}\displaystyle\Pr\{\hat{w}_{i}\geq w_{i}(1-p(1+\tau))|\mathcal{A},|\Upsilon|\leq n/4\} Pr{w^i𝔼[w^i]n(3/43p/2pτ)|𝒜,|Υ|n/4}\displaystyle\leq\Pr\{\hat{w}_{i}-\mathbb{E}[\hat{w}_{i}]\geq n(3/4-3p/2-p\tau)|\mathcal{A},|\Upsilon|\leq n/4\}
e2n(3/43p/2pτ)2\displaystyle\leq e^{-2n(3/4-3p/2-p\tau)^{2}}

where the condition 3/43p/2pτ>03/4-3p/2-p\tau>0 or τ<(3/43p/2)/p\tau<(3/4-3p/2)/p can be satisfied with our choice of free parameter τ\tau since p<1/2p<1/2. Combining everything, we obtain

Pe\displaystyle P_{e} N1clogq+N(e/3)n/4+Ne2n(3/43p/2pτ)2+de2np2τ2\displaystyle\leq N^{1-c\log q}+N(e/3)^{n/4}+Ne^{-2n(3/4-3p/2-p\tau)^{2}}+de^{-2np^{2}\tau^{2}}
N1clogq+N(e/3)n/4+Ne2n(3/43p/2pτ)2+Ne2np2τ2\displaystyle\leq N^{1-c\log q}+N(e/3)^{n/4}+Ne^{-2n(3/4-3p/2-p\tau)^{2}}+Ne^{-2np^{2}\tau^{2}}
=NΩ(logq)+elogNlog(3/e)n/4+2NelogN9/8(0.5p)2n\displaystyle=N^{-\Omega(\log q)}+e^{\log N-\log(3/e)n/4}+2Ne^{\log N-9/8(0.5-p)^{2}n}

where in the last step we pick τ=3(0.5p)4p\tau=\frac{3(0.5-p)}{4p}. Therefore, under the regime d=Ω(log2N)d=\Omega(\log^{2}N), the average probability of error can be bounded as PeNΩ(logq)+3NδP_{e}\leq N^{-\Omega(\log q)}+3N^{-\delta} by choosing n=max{4log(3/e),89(0.5p)2}(1+δ)logNn=\max\{\frac{4}{\log(3/e)},\frac{8}{9(0.5-p)^{2}}\}(1+\delta)\log N. The condition n8dn\leq 8d required in the proof is also satisfied under this regime. Note that the resulting t×Nt\times N binary matrix MM has t=nq=Θ(dlogN)t=nq=\Theta(d\log N) tests.

-C Proof of Theorem 3

We begin with the noiseless case. We will use a recursive approach to obtain an efficiently decodable group testing matrix. Let MnEDM^{\textnormal{ED}}_{n} denote such a matrix with nn columns in the recursion and MnKSM^{\textnormal{KS}}_{n} denote the matrix with nn columns obtained by the Kautz-Singleton construction. Note that the final matrix is MNEDM^{\textnormal{ED}}_{N}. Let tED(d,n,ϵ)t^{\textnormal{ED}}(d,n,\epsilon) and tKS(d,n,ϵ)t^{\textnormal{KS}}(d,n,\epsilon) denote the number of tests for MnEDM^{\textnormal{ED}}_{n} and MnKSM^{\textnormal{KS}}_{n} respectively to detect at most dd defectives among nn columns with average probability of error ϵ\epsilon. We further define DED(d,n,ϵ)D^{\textnormal{ED}}(d,n,\epsilon) to be the decoding time for MnEDM^{\textnormal{ED}}_{n} with tED(d,n,ϵ)t^{\textnormal{ED}}(d,n,\epsilon) rows.

We first consider the case N=d2iN=d^{2^{i}} for some non-negative integer ii. The base case is i=0i=0, i.e., N=dN=d for which we can use individual testing and have tED(d,d,ϵ)=dt^{\textnormal{ED}}(d,d,\epsilon)=d and DED(d,d,ϵ)=O(d)D^{\textnormal{ED}}(d,d,\epsilon)=O(d). For i>0i>0, we use tED(d,N,ϵ/4)×Nt^{\textnormal{ED}}(d,\sqrt{N},\epsilon/4)\times\sqrt{N} matrix MNEDM^{\textnormal{ED}}_{\sqrt{N}} to construct two tED(d,N,ϵ/4)×Nt^{\textnormal{ED}}(d,\sqrt{N},\epsilon/4)\times N matrices M(F)M^{(F)} and M(L)M^{(L)} as follows. The jjth column of MNEDM^{\textnormal{ED}}_{\sqrt{N}} for j[N]j\in[\sqrt{N}] is identical to all iith columns of M(F)M^{(F)} for i[N]i\in[N] if the first 12logN\frac{1}{2}\log N bits of ii is jj where ii and jj are considered as their respective binary representations. Similarly, the jjth column of MNEDM^{\textnormal{ED}}_{\sqrt{N}} for j[N]j\in[\sqrt{N}] is identical to all iith columns of M(L)M^{(L)} for i[N]i\in[N] if the last 12logN\frac{1}{2}\log N bits of ii is jj. We finally construct MNKSM^{\textnormal{KS}}_{N} that achieves ϵ/2\epsilon/2 average probability of error and stack M(F)M^{(F)}, M(L)M^{(L)}, and MNKSM^{\textnormal{KS}}_{N} to obtain the final matrix MNEDM^{\textnormal{ED}}_{N}. Note that, this construction gives us the following recursion in terms of the number of tests

tED(d,N,ϵ)=2tED(d,N,ϵ/4)+tKS(d,N,ϵ/2).\displaystyle t^{\textnormal{ED}}(d,N,\epsilon)=2t^{\textnormal{ED}}(d,\sqrt{N},\epsilon/4)+t^{\textnormal{KS}}(d,N,\epsilon/2).

When N=d2iN=d^{2^{i}}, note that 2i=logdN2^{i}=\log_{d}N and i=loglogdNi=\log\log_{d}N. To solve for tED(d,d2i,ϵ)t^{\textnormal{ED}}(d,d^{2^{i}},\epsilon), we iterate the recursion as follows.

tED(d,d2i,ϵ)\displaystyle t^{\textnormal{ED}}(d,d^{2^{i}},\epsilon) =2tED(d,d2i1,ϵ/4)+tKS(d,d2i,ϵ/2)\displaystyle=2t^{\textnormal{ED}}(d,d^{2^{i-1}},\epsilon/4)+t^{\textnormal{KS}}(d,d^{2^{i}},\epsilon/2)
=4tED(d,d2i2,ϵ/16)+2tKS(d,d2i1,ϵ/8)+tKS(d,d2i,ϵ/2)\displaystyle=4t^{\textnormal{ED}}(d,d^{2^{i-2}},\epsilon/16)+2t^{\textnormal{KS}}(d,d^{2^{i-1}},\epsilon/8)+t^{\textnormal{KS}}(d,d^{2^{i}},\epsilon/2)
\displaystyle\ \ \vdots
=2itED(d,d,ϵ/22i)+j=0i12jtKS(d,d2ij,ϵ/2j+1)\displaystyle=2^{i}t^{\textnormal{ED}}(d,d,\epsilon/2^{2i})+\sum\limits_{j=0}^{i-1}2^{j}t^{\textnormal{KS}}(d,d^{2^{i-j}},\epsilon/2^{j+1})
=2id+j=0i12j4dlog(d2ij/(ϵ/2j+1))\displaystyle=2^{i}\cdot d+\sum\limits_{j=0}^{i-1}2^{j}\cdot 4d\log\left(d^{2^{i-j}}/\left(\epsilon/2^{j+1}\right)\right) (5)
=2id+j=0i12j4d(2ijlogd+(j+1)log2+log(1/ϵ))\displaystyle=2^{i}\cdot d+\sum\limits_{j=0}^{i-1}2^{j}\cdot 4d\left(2^{i-j}\log d+(j+1)\log 2+\log(1/\epsilon)\right)
2id+i2i4dlogd+4dj=0i12j(j+1)+2i4dlog(1/ϵ)\displaystyle\leq 2^{i}\cdot d+i\cdot 2^{i}\cdot 4d\log d+4d\sum\limits_{j=0}^{i-1}2^{j}(j+1)+2^{i}\cdot 4d\log(1/\epsilon)
2id+i2i4dlogd+i2i4d+2i4dlog(1/ϵ)\displaystyle\leq 2^{i}\cdot d+i\cdot 2^{i}\cdot 4d\log d+i\cdot 2^{i}\cdot 4d+2^{i}\cdot 4d\log(1/\epsilon) (6)

where in (5) for simplicity we ignore the term NΩ(logq)N^{-\Omega(\log q)} in the probability of error for Theorem 1 and take tKS(d,N,ϵ)=4dlogN/ϵt^{\textnormal{KS}}(d,N,\epsilon)=4d\log N/\epsilon. Replacing 2i=logdN2^{i}=\log_{d}N and i=loglogdNi=\log\log_{d}N in (6), it follows that

tED(d,N,ϵ)\displaystyle t^{\textnormal{ED}}(d,N,\epsilon) =O(dlogNloglogdN+dlogdNlog((logdN)/ϵ)).\displaystyle=O\left(d\log N\log\log_{d}N+d\log_{d}N\log\left((\log_{d}N)/\epsilon\right)\right).

Note that this gives tED(d,N)=O(dlogNloglogdN)t^{\textnormal{ED}}(d,N)=O(d\log N\log\log_{d}N) in the case where ϵ=Θ(1)\epsilon=\Theta(1).

In the more general case, let i1i\geq 1 be the smallest integer such that d2i1<Nd2id^{2^{i-1}}<N\leq d^{2^{i}}. It follows that i<loglogdN+1i<\log\log_{d}N+1. We can construct MNEDM^{\textnormal{ED}}_{N} from Md2iEDM^{\textnormal{ED}}_{d^{2^{i}}} by removing its last d2iNd^{2^{i}}-N columns. We can operate on MNEDM^{\textnormal{ED}}_{N} as if the removed columns were all defective. Therefore the number of tests satisfies tED(d,N)=O(dlogNloglogdN)t^{\textnormal{ED}}(d,N)=O(d\log N\log\log_{d}N).

We next describe the decoding process. We run the decoding algorithm for MNEDM^{\textnormal{ED}}_{\sqrt{N}} with the components of the outcome vector YY corresponding to M(F)M^{(F)} and M(L)M^{(L)} to compute the estimate sets S^(F)\hat{S}^{(F)} and S^(L)\hat{S}^{(L)}. By induction and the union bound, the set S=S^(F)×S^(L)S^{\prime}=\hat{S}^{(F)}\times\hat{S}^{(L)} contains all the indices iSi\in S with error probability at most ϵ/2\epsilon/2. We further note that |S|d2|S^{\prime}|\leq d^{2}. We finally apply the naive cover decoder to the component of MNEDM^{\textnormal{ED}}_{N} corresponding to MNKSM^{\textnormal{KS}}_{N} over the set SS^{\prime} to compute the final estimate S^\hat{S} which can be done with an additional O(d2tKS(d,N,ϵ/2))O(d^{2}\cdot t^{\textnormal{KS}}(d,N,\epsilon/2)) time. By the union bound overall probability of error is bounded by ϵ\epsilon. This decoding procedure gives us the following recursion in terms of the decoding complexity

DED(d,N,ϵ)=2DED(d,N,ϵ/4)+O(d2tKS(d,N,ϵ/2)).\displaystyle D^{\textnormal{ED}}(d,N,\epsilon)=2D^{\textnormal{ED}}(d,\sqrt{N},\epsilon/4)+O(d^{2}\cdot t^{\textnormal{KS}}(d,N,\epsilon/2)).

When N=d2iN=d^{2^{i}}, to solve for DED(d,d2i,ϵ)D^{\textnormal{ED}}(d,d^{2^{i}},\epsilon), we iterate the recursion as follows.

DED(d,d2i,ϵ)\displaystyle D^{\textnormal{ED}}(d,d^{2^{i}},\epsilon) =2DED(d,d2i1,ϵ/4)+cd2tKS(d,d2i,ϵ/2)\displaystyle=2D^{\textnormal{ED}}(d,d^{2^{i-1}},\epsilon/4)+c\cdot d^{2}\cdot t^{\textnormal{KS}}(d,d^{2^{i}},\epsilon/2)
=4DED(d,d2i2,ϵ/16)+2cd2tKS(d,d2i1,ϵ/8)+cd2tKS(d,d2i,ϵ/2)\displaystyle=4D^{\textnormal{ED}}(d,d^{2^{i-2}},\epsilon/16)+2c\cdot d^{2}\cdot t^{\textnormal{KS}}(d,d^{2^{i-1}},\epsilon/8)+c\cdot d^{2}\cdot t^{\textnormal{KS}}(d,d^{2^{i}},\epsilon/2)
\displaystyle\ \ \vdots
=2iDED(d,d,ϵ/22i)+j=0i12jcd2tKS(d,d2ij,ϵ/2j+1)\displaystyle=2^{i}D^{\textnormal{ED}}(d,d,\epsilon/2^{2i})+\sum\limits_{j=0}^{i-1}2^{j}c\cdot d^{2}\cdot t^{\textnormal{KS}}(d,d^{2^{i-j}},\epsilon/2^{j+1})
=2iO(d)+j=0i12jc4d3log(d2ij/(ϵ/2j+1))\displaystyle=2^{i}\cdot O(d)+\sum\limits_{j=0}^{i-1}2^{j}c\cdot 4d^{3}\log\left(d^{2^{i-j}}/\left(\epsilon/2^{j+1}\right)\right)
2iO(d)+i2i4cd3logd+i2i4cd3+2i4cd3log(1/ϵ)\displaystyle\leq 2^{i}\cdot O(d)+i\cdot 2^{i}\cdot 4cd^{3}\log d+i\cdot 2^{i}\cdot 4cd^{3}+2^{i}\cdot 4cd^{3}\log(1/\epsilon) (7)

where (7) is obtained in the same way as (6). Replacing 2i=logdN2^{i}=\log_{d}N and i=loglogdNi=\log\log_{d}N in (7), it follows that

DED(d,N,ϵ)\displaystyle D^{\textnormal{ED}}(d,N,\epsilon) =O(d3logNloglogdN+d3logdNlog((logdN)/ϵ)).\displaystyle=O\left(d^{3}\log N\log\log_{d}N+d^{3}\log_{d}N\log\left((\log_{d}N)/\epsilon\right)\right).

Note that this gives DED(d,N)=O(d3logNloglogdN)D^{\textnormal{ED}}(d,N)=O(d^{3}\log N\log\log_{d}N) in the case where ϵ=Θ(1)\epsilon=\Theta(1).

The noisy case follows similar lines except the difference is that in the base case where N=dN=d, we cannot use individual testing due to the noise. In this case we can do individual testing with repetitions which requires tED(d,d,ϵ)=O(dlog(d/ϵ))t^{\textnormal{ED}}(d,d,\epsilon)=O(d\log(d/\epsilon)) and DED(d,d,ϵ)=O(dlog(d/ϵ))D^{\textnormal{ED}}(d,d,\epsilon)=O(d\log(d/\epsilon)). We can proceed similarly as in the noiseless case and show that tED(d,N)=O(dlogNloglogdN)t^{\textnormal{ED}}(d,N)=O(d\log N\log\log_{d}N) and DED(d,N)=O(d3logNloglogdN)D^{\textnormal{ED}}(d,N)=O(d^{3}\log N\log\log_{d}N).

Acknowledgements

The third author would like to thank Atri Rudra and Hung Ngo for helpful conversations. We thank the anonymous reviewers for helpful comments and suggestions.

References

  • [1] R. Dorfman, “The detection of defective members of large populations,” The Annals of Mathematical Statistics, vol. 14, no. 4, pp. 436–440, 1943.
  • [2] H-B. Chen and F. K. Hwang, “A survey on nonadaptive group testing algorithms through the angle of decoding,” Journal of Combinatorial Optimization, vol. 15, no. 1, pp. 49–59, 2008.
  • [3] A. Ganesan, S. Jaggi, and V. Saligrama, “Learning immune-defectives graph through group tests,” IEEE Transactions on Information Theory, 2017.
  • [4] D. Malioutov and K. Varshney, “Exact rule learning via boolean compressed sensing,” in International Conference on Machine Learning, 2013, pp. 765–773.
  • [5] A. C. Gilbert, M. A. Iwen, and M. J. Strauss, “Group testing and sparse signal recovery,” in Signals, Systems and Computers, 2008 42nd Asilomar Conference on. IEEE, 2008, pp. 1059–1063.
  • [6] A. Emad and O. Milenkovic, “Poisson group testing: A probabilistic model for nonadaptive streaming boolean compressed sensing,” in Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on. IEEE, 2014, pp. 3335–3339.
  • [7] T. Berger, N. Mehravari, D. Towsley, and J. Wolf, “Random multiple-access communication and group testing,” Communications, IEEE Transactions on, vol. 32, no. 7, pp. 769 – 779, jul 1984.
  • [8] J. K. Wolf, “Born again group testing: Multiaccess communications,” Information Theory, IEEE Transactions on, vol. 31, no. 2, pp. 185–191, 1985.
  • [9] J. Luo and D. Guo, “Neighbor discovery in wireless ad hoc networks based on group testing,” in Communication, Control, and Computing, 2008 46th Annual Allerton Conference on. IEEE, 2008, pp. 791–797.
  • [10] A. K. Fletcher, V.K. Goyal, and S. Rangan, “A sparsity detection framework for on-off random access channels,” in Information Theory, 2009. ISIT 2009. IEEE International Symposium on. IEEE, 2009, pp. 169–173.
  • [11] H. Q. Ngo and D-Z. Du, “A survey on combinatorial group testing algorithms with applications to dna library screening,” Discrete mathematical problems with medical applications, vol. 55, pp. 171–182, 2000.
  • [12] D. Du and F. K. Hwang, Combinatorial group testing and its applications, vol. 12, World Scientific, 2000.
  • [13] G. K. Atia and V. Saligrama, “Boolean compressed sensing and noisy group testing,” Information Theory, IEEE Transactions on, vol. 58, no. 3, pp. 1880–1901, 2012.
  • [14] D. Sejdinovic and O. Johnson, “Note on noisy group testing: Asymptotic bounds and belief propagation reconstruction,” CoRR, vol. abs/1010.2441, 2010.
  • [15] C. L. Chan, P. H. Che, S. Jaggi, and V. Saligrama, “Non-adaptive probabilistic group testing with noisy measurements: Near-optimal bounds with efficient algorithms,” in 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Sept 2011, pp. 1832–1839.
  • [16] J. Scarlett and V. Cevher, “Phase Transitions in Group Testing,” in ACM-SIAM Symposium on Discrete Algorithms (SODA), 2016.
  • [17] M. Aldridge, L. Baldassini, and O. Johnson, “Group testing algorithms: Bounds and simulations,” IEEE Transactions on Information Theory, vol. 60, no. 6, pp. 3671–3687, June 2014.
  • [18] O. Johnson, M. Aldridge, and J. Scarlett, “Performance of group testing algorithms with near-constant tests per item,” IEEE Transactions on Information Theory, vol. 65, no. 2, pp. 707–723, Feb 2019.
  • [19] J. Scarlett and V. Cevher, “Near-optimal noisy group testing via separate decoding of items,” IEEE Journal of Selected Topics in Signal Processing, vol. 12, no. 5, pp. 902–915, Oct 2018.
  • [20] J. Scarlett and O. Johnson, “Noisy Non-Adaptive Group Testing: A (Near-)Definite Defectives Approach,” ArXiv e-prints, Aug. 2018.
  • [21] A. G. D’yachkov and V. V. Rykov, “Bounds on the length of disjunctive codes,” Problemy Peredachi Informatsii, vol. 18, no. 3, pp. 7–13, 1982.
  • [22] Z. Furedi, “On r-cover-free families,” Journal of Combinatorial Theory, Series A, vol. 73, no. 1, pp. 172–173, 1996.
  • [23] W. Kautz and R. Singleton, “Nonrandom binary superimposed codes,” IEEE Transactions on Information Theory, vol. 10, no. 4, pp. 363–377, October 1964.
  • [24] E. Porat and A. Rothschild, “Explicit non-adaptive combinatorial group testing schemes,” Automata, Languages and Programming, pp. 748–759, 2008.
  • [25] A. Mazumdar, “Nonadaptive group testing with random set of defectives,” IEEE Transactions on Information Theory, vol. 62, no. 12, pp. 7522–7531, Dec 2016.
  • [26] H. Q. Ngo, E. Porat, and A. Rudra, “Efficiently decodable error-correcting list disjunct matrices and applications,” in Automata, Languages and Programming, Berlin, Heidelberg, 2011, pp. 557–568, Springer Berlin Heidelberg.
  • [27] A. J. Macula, V. V. Rykov, and S. Yekhanin, “Trivial two-stage group testing for complexes using almost disjunct matrices,” Discrete Applied Mathematics, vol. 137, no. 1, pp. 97 – 107, 2004.
  • [28] M. B. Malyutov, “The separating property of random matrices,” Math. Notes, vol. 23, no. 1, pp. 84––91, 1978.
  • [29] A. Zhigljavsky, “Probabilistic existence theorems in group testing,” J. Statist. Planning Inference, vol. 115, no. 1, pp. 1–43, 2003.
  • [30] I. S. Reed and G. Solomon, “Polynomial codes over certain finite fields,” Journal of the Society for Industrial and Applied Mathematics, vol. 8, no. 2, pp. 300–304, 1960.
  • [31] Y. Erlich, “Combinatorial pooling using RS codes,” https://github.com/TeamErlich/pooling, 2017.
  • [32] Y. Erlich, A. Gilbert, H. Ngo, A. Rudra, N. Thierry-Mieg, M. Wootters, D. Zielinski, and O. Zuk, “Biological screens from linear codes: theory and tools,” bioRxiv, 2015.
  • [33] M. Cheraghchi, “Noise-resilient group testing: Limitations and constructions,” in Fundamentals of Computation Theory, Berlin, Heidelberg, 2009, pp. 62–73, Springer Berlin Heidelberg.
  • [34] P. Indyk, H. Q. Ngo, and A. Rudra, “Efficiently decodable non-adaptive group testing,” in Proceedings of the Twenty-First Annual ACM-SIAM Symposium on Discrete Algorithms. Society for Industrial and Applied Mathematics, 2010, pp. 1126–1142.
  • [35] K. Lee, R. Pedarsani, and K. Ramchandran, “Saffron: A fast, efficient, and robust framework for group testing based on sparse-graph codes,” in Information Theory (ISIT), 2016 IEEE International Symposium on. IEEE, 2016, pp. 2873–2877.
  • [36] S. Cai, M. Jahangoshahi, M. Bakshi, and S. Jaggi, “Efficient algorithms for noisy group testing,” IEEE Transactions on Information Theory, vol. 63, no. 4, pp. 2113–2136, April 2017.
  • [37] C. L. Chan, S. Jaggi, V. Saligrama, and S. Agnihotri, “Non-adaptive group testing: Explicit bounds and novel algorithms,” IEEE Transactions on Information Theory, vol. 60, no. 5, pp. 3019–3035, May 2014.
  • [38] A. G. D’yachkov and V. V. Rykov, “A survey of superimposed code theory,” Problems Control Inform. Theory/Problemy Upravlen. Teor. Inform., vol. 12, pp. 229–242, 1983.
  • [39] A. De Bonis., L. Gasieniec, and U. Vaccaro, “Optimal two-stage algorithms for group testing problems,” SIAM J. Comput., vol. 34, no. 5, pp. 1253–1270, 2005.
  • [40] A. M. Rashad, “Random coding bounds on the rate for list-decoding superimposed codes,” Problems Control Inform. Theory/Problemy Upravlen. Teor. Inform., vol. 19(2), pp. 141–149, 1990.
  • [41] M. Sobel, S. Kumar, and S. Blumenthal, “Symmetric binomial group-testing with three outcomes,” in Proc. Purdue Symp. Decision Procedure, 1971, pp. 119–160.
  • [42] F. Hwang, “Three versions of a group testing game,” SIAM Journal on Algebraic Discrete Methods, vol. 5, no. 2, pp. 145–153, 1984.
  • [43] H-B. Chen and H-L. Fu, “Nonadaptive algorithms for threshold group testing,” Discrete Applied Mathematics, vol. 157, no. 7, pp. 1581 – 1585, 2009.
  • [44] M. Cheraghchi, “Improved constructions for non-adaptive threshold group testing,” CoRR, vol. abs/1002.2244, 2010.
  • [45] T. V. Bui, M. Kuribayashi, M. Cheraghchi, and I. Echizen, “Efficiently decodable non-adaptive threshold group testing,” CoRR, vol. abs/1712.07509, 2017.
  • [46] H. A. Inan, P. Kairouz, and A. Ozgur, “Sparse group testing codes for low-energy massive random access,” in 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Oct 2017, pp. 658–665.
  • [47] V. Gandikota, E. Grigorescu, S. Jaggi, and S. Zhou, “Nearly optimal sparse group testing,” in 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Sept 2016, pp. 401–408.
  • [48] H. A. Inan, P. Kairouz, and A. Ozgur, “Energy-limited massive random access via noisy group testing,” in Information Theory (ISIT), 2018 IEEE International Symposium on. IEEE, 2018, pp. 1101–1105.
  • [49] T. Hartman and R. Raz, “On the distribution of the number of roots of polynomials and explicit weak designs,” Random Structures & Algorithms, vol. 23, no. 3, pp. 235–263, 2003.
  • [50] R. Bardenet and O-A. Maillard, “Concentration inequalities for sampling without replacement,” Bernoulli, vol. 21, no. 3, pp. 1361–1385, 08 2015.