This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Interfacing biology, category theory and mathematical statistics

Dominique Pastor1{\>}^{1}   Erwan Beurier1{\>}^{1} 1   IMT Atlantique
LabSTICC & LEGO
Université de Bretagne-Loire
29238 Brest, France { erwan.beurier ; dominique.pastor ; roger.waldeck }@imt-atlantique.fr 2   Faculté des Sciences
Dépt. de Mathématiques
LAMFA
Université de Picardie Jules Verne
33 rue Saint-Leu
F-80039 Amiens, France
   Andrée Ehresmann2{\>}^{2}     Roger Waldeck1{\>}^{1} 2   Faculté des Sciences
Dépt. de Mathématiques
LAMFA
Université de Picardie Jules Verne
33 rue Saint-Leu
F-80039 Amiens, France andree.ehresmann@u-picardie.fr
Abstract

Motivated by the concept of degeneracy in biology [4], we establish a first connection between the Multiplicity Principle [5, 6] and mathematical statistics. Specifically, we exhibit two families of tests that satisfy this principle to achieve the detection of a signal in noise.

1 Introduction

In [4], Edelman & Gally pointed out degeneracy as the fundamental property allowing for living systems to evolve through natural selection towards more complexity in fluctuating environments. Degeneracy is defined [4] as “ … the ability of elements that are structurally different to perform the same function or yield the same output”. Degeneracy is a crucial feature of immune systems and neural networks, at all organization levels.

The Multiplicity Principle (MP) [5, 6], introduced by Ehresmann & Vanbremeersch, is a mathematical formalization of degeneracy in Categorical terms. The consequences of this principle, as treated in [5, 6], underpin Edelman & Gally’s conjecture according to which “complexity and degeneracy go hand in hand[4].

Another property of many biological and social systems is their resilience: (i) they can perform in degraded mode, with some performance loss, but without collapsing; (ii) they can recover their initial performance level when nominal conditions are satisfied again; (iii) they can perform corrections and auto-adaption so as to maintain essential tasks for their survival. In addition, resilience of social or biological systems is achieved via agents with different skills. For instance, cells are simply reactive organisms, whereas social agents have some cognitive properties. Thence the idea that resilience may derive from fundamental properties satisfied by agents, interactions and organizations. Could this fundamental property be a possible consequence of degeneracy [6, Section 3.1, p. 15]?

The notion of resilience remains, however, somewhat elusive, mathematically speaking. In contrast, the notion of robustness has a long history and track record in mathematical statistics [7]. By and large, a statistical method is robust if its performance is not unduly altered in case of outliers or fluctuations around the model for which it is designed. Can we fathom the links between resilience and robustness?

As an attempt to embrace the questions raised above from a comprehensive outlook, the original question addressed in this work is the possible connection between MP and robustness to account for emergence of resilience in complex systems. As a first step in our study aimed at casting the notions of robustness, resilience and degeneracy within the same theoretical framework based on the MP, we hereafter establish that statistical tests do satisfy the MP. The task to perform by the tests is the fundamental problem of detecting a signal in noise. However, to ease the reading of a paper at the interface between category theory and mathematical statistics, we consider a simplified version of this problem.

The paper is organized as follows. We begin by specifying notation and notions in mathematical statistics. In Section 2, we state the MP in categorical words on the basis of [5] and consider the particular case of preorders, which will be sufficient at the present time to establish that statistical tests satisfy the MP for detecting signals in noise. In Section 3, we set out the statistical detection problem. We will then introduce, in Section 4, a preorder that makes it possible to exhibit two types of "structurally different" tests, namely, the Neyman-Pearson tests (Section 5) and the RDT tests (Section 6). Section 7 concludes the paper by establishing that these two types of tests achieve the MP for the detection problem under consideration. For space considerations, we limit proofs to the minimum making it possible to follow the approach without too much undue effort.

Summary of main results

Because this paper lies at the interface between different mathematical specialties, the present section summarizes its contents in straight text. To begin with, the MP is a property that a category may satisfy when it involves structurally different diagrams sharing the same cocones. To state our main results, it will not be necessary to consider the general MP though. In fact, the particular case of preordered sets will suffice, in which case the MP reduces to Proposition 1.
Second, in statistical hypothesis testing, a hypothesis can be seen as a predicate, of which we can aim at determining the truth value by using statistical decisions. There exist many optimality criteria to devise a decision to test a given hypothesis. In non-Bayesian approaches, which will be our focus below, such criteria are specified through the notions of size and power.

The size is the least upper bound for the probability of rejecting the hypothesis when this one is actually true. We generally want this size to remain below a certain value called level, because the hypothesis to test mostly represents the standard situation. For instance, planes in the sky are rare events, after all, and the standard hypothesis is "there is no plane", which represents the nominal situation. A too large level may result in an intolerable cluttering of a radar screen.

We do not want to be bothered by too many alarms. In contrast, when the hypothesis is false, we want to reject it with the highest possible confidence. The probability that a decision rejects the hypothesis when this one is actually false is called the power of the decision. For a given testing problem, we thus look for decisions with maximal power within the set of those decisions that have a size less than or equal to a a specified level. This defines a preorder. A maximal element in this preorder is said to be optimal.

Different hypotheses to test may thus require different criteria, specified through different notions of size and different notions of power. This is what we exploit below to exhibit two sets of "structurally different" decisions that satisfy the MP.

To carry out this construction, we consider the detection of a signal in independent standard gaussian noise, a classical problem in many applications. This is an hypothesis testing problem for which there exists an optimality criterion where the size is the so-called probability of false alarm and the power is the so-called probability of detection. This criterion has a solution, the Neyman-Pearson (NP) decision, which is thus the maximal element of a certain preorder. We can consider a second class of decisions, namely, the RDT decisions. These decisions are aimed at detecting deviations of a signal with respect to a known deterministic model in presence of independent standard gaussian noise. This problem is rotationally invariant and the RDT decisions are optimal with respect to a specific criterion defined through suitable notions of size and power. They are maximal elements of another preordered set. Although not dedicated to signal detection, these decisions can be used as surrogates to NP decisions to detect a signal. It turns out that the family of RDT decisions and that of NP decisions satisfy the MP as stated in Theorem 4. This is because the more data we have, the closer to perfection both decisions are.

Notation

Random variables. Given two measurable spaces \mathcal{E} and \mathcal{F}, (,)\mathcal{M}(\mathcal{E},\mathcal{F}) denotes the set of all measurable functions defined on \mathcal{E} and valued in \mathcal{F}. The two σ\sigma-algebra involved are omitted in the notation because, in the sequel, they will always be obvious from the context. In particular, we will throughout consider a probability space (Ω,,)(\Omega,\mathcal{B},\mathbb{P}) and systematically endow \mathbb{R} with the Borel σ\sigma-algebra, which will not be recalled. Therefore, (Ω,)\mathcal{M}(\Omega,\mathbb{R}) designates the set of all real random variables and (Ω,n)\mathcal{M}(\Omega,\mathbb{R}^{n}) is the set of nn-dimensional real random vectors.

Given q[0,[q\in[0,\infty[, (q)\mathcal{B}_{\infty}(q) is the set of all real random variables Δ(Ω,)\Delta\in\mathcal{M}(\Omega,\mathbb{R}) such that |Δ|_q{\left|\Delta\right|_{\_}{\infty}}\leqslant q. As usual, we write X𝒩(0,1)X\thicksim\mathcal{N}(0,1) to mean that X(Ω,)X\in\mathcal{M}(\Omega,\mathbb{R}) is standard normal. Given a sequence (X_n)_n(Ω,)(X_{\_}n)_{\_}{n\in\mathbb{N}}\in\mathcal{M}(\Omega,\mathbb{R})^{\mathbb{N}} of real random variables, we write X_1,X_2,iid𝒩(0,1)X_{\_}1,X_{\_}2,\ldots\stackrel{{\scriptstyle\text{iid}}}{{\thicksim}}\mathcal{N}(0,1) to mean that X_1,X_2,X_{\_}1,X_{\_}2,\ldots are independent and identically distributed with common distribution 𝒩(0,1)\mathcal{N}(0,1).
Decisions et Observations. Throughout, ({0,1}×Ω,{0,1})\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right) designates the set of all measurable functions D:{0,1}×Ω{0,1}{D}:\big{\{}0,1\big{\}}\times\Omega\to\big{\{}0,1\big{\}}. Any element of ({0,1}×Ω,{0,1})\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right) is called a decision for obvious reasons given below. If D({0,1}×Ω,{0,1}){D}\in\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right) then, for any ε{0,1}\varepsilon\in\big{\{}0,1\big{\}}, D(ε){D}(\varepsilon) denotes the Bernoulli-distributed random variable D(ε):Ω{0,1}{D}(\varepsilon):\Omega\to\big{\{}0,1\big{\}} defined for any given ωΩ\omega\in\Omega by D(ε)(ω)=D(ε,ω){D}(\varepsilon)(\omega)={D}(\varepsilon,\omega). An nn-dimensional test is hereafter any measurable function f:n{0,1}f:\mathbb{R}^{n}\to\{0,1\} and (n,{0,1})\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) stands for the set of all nn-dimensional tests. A measurable function X:{0,1}×ΩnX:\left\{0,1\right\}\times\Omega\to\mathbb{R}^{n} is hereafter called an observation and ({0,1}×Ω,n)\mathcal{M}\left(\{0,1\}\times\Omega,\mathbb{R}^{n}\right) denotes the set of all these observations. Given a test f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) and X({0,1}×Ω,n)X\in\mathcal{M}\left(\{0,1\}\times\Omega,\mathbb{R}^{n}\right), D=f(X){D}=f(X) is trivially a decision: D({0,1}×Ω,{0,1}){D}\in\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right). If X({0,1}×Ω,n)X\in\mathcal{M}\left(\{0,1\}\times\Omega,\mathbb{R}^{n}\right) then, for any ε{0,1}\varepsilon\in\{0,1\}, X(ε)=X(ε,)(Ω,n)X(\varepsilon)=X(\varepsilon,\cdot)\in\mathcal{M}(\Omega,\mathbb{R}^{n}) is defined for every ωΩ\omega\in\Omega by X(ε)(ω)=X(ε,ω)X(\varepsilon)(\omega)=X(\varepsilon,\omega).
Empirical means. We define the empirical mean of a given sequence y=(y_n)_ny=(y_{\_}n)_{\_}{n\in\mathbb{N}} of real values as the sequence (yn)_n\left(\langle{y}\rangle_{n}\right)_{\_}{n\in\mathbb{N}} of real values such that, n,yn:=1n_i=1ny_i\forall n\in\mathbb{N},\langle{y}\rangle_{n}\,\,:=\,\,\frac{1}{n}\sum_{\_}{i=1}^{n}y_{\_}i. By extension, the empirical mean of a sequence Y=(Y_n)_nY=(Y_{\_}n)_{\_}{n\in\mathbb{N}} of random variables where each Y_n(Ω,)Y_{\_}n\in\mathcal{M}(\Omega,\mathbb{R}) is the sequence (Yn)_n\left(\langle{Y}\rangle_{n}\right)_{\_}{n\in\mathbb{N}} of random variables where, for each nn\in\mathbb{N}, Yn(Ω,)\langle{Y}\rangle_{n}\in\mathcal{M}(\Omega,\mathbb{R}) is defined by Yn:=1n_i=1nY_i\langle{Y}\rangle_{n}\,\,:=\,\,\frac{1}{n}\sum_{\_}{i=1}^{n}Y_{\_}i. Therefore, for any ωΩ\omega\in\Omega, Yn(ω):=Y(ω)n\langle{Y}\rangle_{n}(\omega)\,\,:=\,\,\langle{Y(\omega)}\rangle_{n} with Y(ω)=(Y_n(ω))_nY(\omega)=(Y_{\_}n(\omega))_{\_}{n\in\mathbb{N}}. If Y=(Y_n)_nY=(Y_{\_}n)_{\_}{n\in\mathbb{N}} is a sequence of observations (n\forall\,n\in\mathbb{N}, Y_n({0,1}×Ω,)Y_{\_}n\in\mathcal{M}\left(\{0,1\}\times\Omega,\mathbb{R}\right)), we define the empirical mean of YY as the sequence (Yn)_n\left(\langle{Y}\rangle_{n}\right)_{\_}{n\in\mathbb{N}} of observations such that, for ε{0,1}\varepsilon\in\{0,1\}, Yn({0,1}×Ω,)\langle{Y}\rangle_{n}\in\mathcal{M}\left(\{0,1\}\times\Omega,\mathbb{R}\right) with Yn(ε)=Y(ε)n\langle{Y}\rangle_{n}(\varepsilon)=\langle{Y(\varepsilon)}\rangle_{n} and Y(ε)=(Y_n(ε))_nY(\varepsilon)=(Y_{\_}n(\varepsilon))_{\_}{n\in\mathbb{N}}.
Preordered sets. Given a preordered set (E,)(E,\preceq) and AEA\subset E, the set of maximal elements of AA is denoted by max(A,(E,))\max\left(A,(E,\preceq)\right), the set of upper bounds of AA is denoted by upper(A,(E,))\textup{upper}\left(A,(E,\preceq)\right) and the set of least upper bounds of AA in (E,)(E,\preceq) is denoted by sup(A,(E,))\sup\left(A,(E,\preceq)\right).

2 Multiplicity Principle

2.1 General case

The multiplicity principle (MP) comes from [5]. It proposes a categorical approach to the biological degeneracy principle, which ensures a kind of flexible redundancy. Roughly, MP, in a category 𝒞\mathscr{C}, ensures the existence of structurally non isomorphic diagrams with same colimit. A formal definition relies on the notion of a cluster between diagrams in a category 𝒞\mathscr{C}.

Definition 1 (Cluster).

Let D:𝒟𝒞D:\mathscr{D}\to\mathscr{C} and E:𝒞E:\mathscr{E}\to\mathscr{C} be two (small) diagrams. A cluster G:DEG:D\to E is a maximal set G={f:D(d)E(e)|d𝒟,e,f𝒞}G=\left\{f:D(d)\to E(e)\ |\ d\in\mathscr{D},e\in\mathscr{E},f\in\mathscr{C}\right\} such that:

  1. (i)

    for all d𝒟d\in\mathscr{D} there exist ee\in\mathscr{E} and g:D(d)E(e)g:D(d)\to E(e) such that gGg\in G

  2. (ii)

    let G(d)G(d) be the subset of GG consisting of arrows g:D(d)E(e)g:D(d)\to E(e) associated to the same dd; then G(d)G(d) is included in a connected component of the comma-category (D(d)|E)\left(D(d)\ |\ E\right)

  3. (iii)

    if g:D(d)E(e)G(d)g:D(d)\to E(e)\in G(d) and ε:ee\varepsilon:e\to e^{\prime}\in\mathscr{E}, then E(ε)gG(d)E(\varepsilon)\circ g\in G(d)

  4. (iv)

    if δ:dd𝒟\delta:d^{\prime}\to d\in\mathscr{D} and g:D(d)E(e)G(d)g:D(d)\to E(e)\in G(d), then gD(δ)G(d)g\circ D(\delta)\in G(d^{\prime})

For instance, a connected cone from cc to DD can be seen as a cluster from the constant functor Δ(c)\Delta(c) to DD; and any cocone from EE to cc is a cluster EΔ(c)E\to\Delta(c).

Remark 1.

Adjacent clusters can be composed: a cluster G:DE_0G:D\to E_{\_}0 and a cluster G_0:E_0EG_{\_}0:E_{\_}0\to E can be composed to a cluster G_0GG_{\_}0\circ G. We can then consider a category of clusters of 𝒞\mathscr{C}, whose objects are the (small) diagrams 𝒟𝒞\mathscr{D}\to\mathscr{C}, and an arrow DED\to E is a cluster. This category is isomorphic to the free cocompletion of 𝒞\mathscr{C} [5].

A cluster G:DEG:D\to E defines a functor ΩG:𝐂𝐨𝐜𝐨𝐧𝐞𝐬(E)𝐂𝐨𝐜𝐨𝐧𝐞𝐬(D)\Omega G:\mathbf{Cocones}\left(E\right)\to\mathbf{Cocones}\left(D\right) mapping a cocone α\alpha to the cocone αG\alpha\circ G (composite of α\alpha, seen as a cluster, and GG, which is a cluster).

Definition 2 (Multiplicity principle (MP)).

A category  𝒞\mathscr{C} satisfies the multiplicity principle (MP) if there exist two diagrams D:𝒟𝒞D:\mathscr{D}\to\mathscr{C} and E:𝒞E:\mathscr{E}\to\mathscr{C} such that:

  1. (i)

    𝐂𝐨𝐜𝐨𝐧𝐞𝐬(D)𝐂𝐨𝐜𝐨𝐧𝐞𝐬(E)\mathbf{Cocones}\left(D\right)\cong\mathbf{Cocones}\left(E\right);

  2. (ii)

    There is no cluster G:DEG:D\to E nor G:EDG:E\to D such that ΩG\Omega G is an isomorphism.

DD and EE having the same cocones translates the property of both systems to accomplish the same function. The absence of clusters between DD and EE that define an isomorphism, reflects the structural difference between DD and EE, which is key to robustness and adaptability: if the system described by EE fails, then DD may replace it.

2.2 Application to preorders

The main purpose of this paper is to find a meaningful instance of the MP in some preorder. In the following, we do not distinguish between a preorder and its associated category.

Proposition 1 (MP in a preorder).

Let (E,)\left(E,\leqslant\right) be a preorder. If there are two disjoint subsets A,BEA,B\subset E such that the following conditions hold, then EE verifies the MP:

  1. (i)

    AA and BB have the same sets of upper bounds

  2. (ii)

    There is an aAa\in A with no upper bounds in BB

  3. (iii)

    There is a bBb\in B with no upper bounds in AA

Proof.

Condition (i) ensures that AA and BB have isomorphic categories of cocones. Conditions (ii) and (iii) respectively ensure that there is no cluster i_Ai_Bi_{\_}{A}\to i_{\_}{B} nor i_Bi_Ai_{\_}{B}\to i_{\_}{A} where i_A:AEi_{\_}{A}:A\hookrightarrow E and i_B:BEi_{\_}{B}:B\hookrightarrow E are the inclusion functors. ∎

Albeit trivial, the following lemma will be helpful.

Lemma 1.

Given a preordered set (E,)(E,\preceq), if AA and BB are two subsets of EE such that A×B=\,A\times\!B\,\,\cap\preceq\,\,=\,\emptyset and sup(A,(E,))=sup(B,(E,))\sup\left(A,(E,\preceq)\right)=\sup\left(B,(E,\preceq)\right), then EE satisfies the MP.

3 Statistical detection of a signal in noise

3.1 Problem statement

Let ε{0,1}\varepsilon\in\{0,1\} be the unknown indicator value on whether a certain physical phenomenon has occurred (ε=1\varepsilon=1) or not (ε=0\varepsilon=0). We aim at determining this value. It is desirable to resort to something more evolved than tossing a coin to estimate ε\varepsilon. However, whatever D{D}, the decision is erroneous for any ωΩ\omega\in\Omega such that D(ε,ω)ε{D}(\varepsilon,\omega)\neq\varepsilon. We thus have two distinct cases.
False alarm probability: If ε=0\varepsilon=0 and D(0,ω)=1{D}(0,\omega)=1, we commit a false alarm or error of the 1st1^{\text{st}} kind, since we have erroneously decided that the phenomenon has occurred while nothing actually happened. We thus define the false alarm probability (aka size, aka error probability of the 1st1^{\text{st}} kind) of D{D} as:

fa[D]=def[D(0)=1]\mathbb{P}_{\textsc{fa}}\left[{{D}}\right]\stackrel{{\scriptstyle\text{def}}}{{=}}\mathbb{P}\big{[}{D}(0)=1\big{]} (1)

Detection probability: If ε=1\varepsilon=1 and D(1,ω)=0{D}(1,\omega)=0, we commit an error of the 2nd2^{\text{nd}} kind, also called missed detection since, in this case, we have missed the occurrence of the phenomenon. As often in the literature on the topic, we prefer to use the probability of correctly detecting the phenomenon and we define the detection probability as:

det[D]=def[D(1)=1]\mathbb{P}_{\textsc{det}}\left[{{D}}\right]\stackrel{{\scriptstyle\text{def}}}{{=}}\mathbb{P}\big{[}{D}(1)=1\big{]} (2)

3.2 Decision with level γ(0,1)\gamma\in(0,1) and oracles

Among all possible decisions, the omniscient oracle D({0,1}×Ω,{0,1}){D}^{*}\in\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right) is defined for any pair (ε,ω){0,1}×Ω(\varepsilon,\omega)\in\big{\{}0,1\big{\}}\times\Omega by setting D(ε,ω)=ε{D}^{*}(\varepsilon,\omega)=\varepsilon. Its probability of false alarm is 0 and its probability of detection is 11: fa[D]=0\mathbb{P}_{\textsc{fa}}\left[{{D}^{*}}\right]=0 et det[D]=1\mathbb{P}_{\textsc{det}}\left[{{D}^{*}}\right]=1. This omniscient oracle has no practical interest since it knows ε\varepsilon. That’s not really fair! Since it is not possible in practice to guarantee a null false alarm probability, we focus on decisions whose false alarm probabilities are upper-bounded by a real number γ(0,1)\gamma\in(0,1) called level. We state the following definition.

Definition 3 (Level).

Given γ]0,1[\gamma\in]0,1[, we say that D({0,1}×Ω,{0,1}){D}\in\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right) has level γ\gamma if fa[D]γ\,\mathbb{P}_{\textsc{fa}}\left[{{D}}\right]\leqslant\gamma. The set of all decisions with level γ]0,1[\gamma\in]0,1[ is denoted by Decγ\mathrm{Dec}_{\gamma}.

We can easily prove the existence of an infinite number of elements in Decγ\mathrm{Dec}_{\gamma} that all have a detection probability equal to 11. Whence the following definition.

Definition 4.

Given γ(0,1)\gamma\in(0,1), an oracle with level γ\gamma is any decision DDecγ{D}\in\mathrm{Dec}_{\gamma} such that det[D]=1\mathbb{P}_{\textsc{det}}\left[{{D}}\right]=1. The set of all the oracles with level γ\gamma is denoted by 𝕆γ\mathbb{O}_{\gamma}.

Oracles with level γ\gamma have no practical interest either since they require prior knowledge of ε\varepsilon! Therefore, we restrict our attention to decisions in Decγ\mathrm{Dec}_{\gamma} that "approximate" at best the oracles with level γ\gamma, without prior knowledge of ε\varepsilon, of course. To this end, we must preorder decisions.

Lemma-Definition 1 (Total preorder (Decγ,)(\mathrm{Dec}_{\gamma},\,\,{\preceq}\,\,)).

For any given γ(0,1)\gamma\in(0,1) and any pair (D,D)Decγ×Decγ({D},{D}^{\prime})\in\mathrm{Dec}_{\gamma}\times\mathrm{Dec}_{\gamma}, we define a preorder (Decγ,)(\mathrm{Dec}_{\gamma},\,\,{\preceq}\,\,) by setting:

DDifdet[D]det[D].{D}\,\,{\preceq}\,\,{D}^{\prime}\quad\text{if}\quad\mathbb{P}_{\textsc{det}}\left[{{D}}\right]\leqslant\mathbb{P}_{\textsc{det}}\left[{{D}^{\prime}}\right]. (3)

We write DD{D}\cong{D}^{\prime} if DD{D}\,\,{\preceq}\,\,{D}^{\prime} and DD{D}^{\prime}\,\,{\preceq}\,\,{D}.

3.3 Observations

In practice, observations help us decide whether the phenomenon has occurred or not. By collecting a certain number of them, we can expect to make a decision. Hereafter, observations are assumed to be elements of ({0,1}×Ω,)\mathcal{M}\left(\{0,1\}\times\Omega,\mathbb{R}\right) and corrupted versions of ε\varepsilon. We suppose that we have a sequence (Y_n)_n(Y_{\_}n)_{\_}{n\in\mathbb{N}} of such random variables. As a first standard model, we could assume that, for any nn\in\mathbb{N} and any (ε,ω){0,1}×Ω(\varepsilon,\omega)\in\{0,1\}\times\Omega, Y_n(ε,ω)=ε+X_n(ω)Y_{\_}n(\varepsilon,\omega)=\varepsilon+X_{\_}n(\omega) with X_1,X_2,,X_n,iid𝒩(0,1)X_{\_}1,X_{\_}2,\ldots,X_{\_}n,\ldots\stackrel{{\scriptstyle\text{iid}}}{{\thicksim}}\mathcal{N}(0,1). In this additive model, X_nX_{\_}n models noise on the nnth observation. We could make this model more complicated and realistic by considering random vectors instead of variables. However, with respect to our purpose, the significant improvement we can bring to the model is elsewhere. Indeed, we have assumed above that the signal, regardless of noise, is ε\varepsilon. However, from a practical point of view, it is more realistic to assume that the nnth observation Y_nY_{\_}n captures ε\varepsilon in presence of some interference Δ_n\Delta_{\_}n, independent of X_nX_{\_}n. In practice, the probability distribution of Δ_n\Delta_{\_}n will hardly be known and, as a means to compensate for this lack of knowledge, we assume the existence of a uniform bound on the amplitude of all possible interferences. Therefore, we assume that, for all (ε,ω){0,1}×Ω(\varepsilon,\omega)\in\{0,1\}\times\Omega, Y_n(ε,ω)=ε+X_n(ω)+Δ_n(ω)Y_{\_}n(\varepsilon,\omega)=\varepsilon+X_{\_}n(\omega)+\Delta_{\_}n(\omega) and the existence of q[0,)q\in[0,\infty) such that Δ_n(q)\Delta_{\_}n\in\mathcal{B}_{\infty}(q). After all, this model is standard in time series analysis: ε\varepsilon plays the role of a trend, Δ_n\Delta_{\_}n is the seasonal variation and X_nX_{\_}n is the measurement noise.

For each q[0,)q\in[0,\infty), Seqq\textup{Seq}_{\,q} henceforth designates the set of all the sequences:

Y=(Y_n)_n({0,1}×Ω,{0,1}){Y}=\left(Y_{\_}n\right)_{\_}{n\in\mathbb{N}}\in\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right)^{\!\mathbb{N}}

such that, n\forall n\in\mathbb{N} and (ε,ω){0,1}×Ω\forall(\varepsilon,\omega)\in\{0,1\}\times\Omega, Y_n(ε,ω)=ε+Δ_n(ω)+X_n(ω)Y_{\_}n(\varepsilon,\omega)=\varepsilon+\Delta_{\_}n(\omega)+X_{\_}n(\omega), where Δ_n(q)\Delta_{\_}n\in\mathcal{B}_{\infty}(q) and X_n𝒩(0,1)X_{\_}n\thicksim\mathcal{N}(0,1) are independant. Therefore, for all nn\in\mathbb{N} and all ε{0,1}\varepsilon\in\{0,1\}, Y_n(ε)=ε+Δ_n+X_nY_{\_}n(\varepsilon)=\varepsilon+\Delta_{\_}n+X_{\_}n, with X_1,X_2,,X_n,iid𝒩(0,1)X_{\_}1,X_{\_}2,\ldots,X_{\_}n,\ldots\stackrel{{\scriptstyle\text{iid}}}{{\thicksim}}\mathcal{N}(0,1).

4 Selectivity, landscapes of tests and preordering

For any sequence Y=(Y_n)_n({0,1}×Ω,{0,1}){Y}=\left(Y_{\_}n\right)_{\_}{n\in\mathbb{N}}\in\mathcal{M}\!\left(\{0,1\}\!\times\!\Omega,\big{\{}0,1\big{\}}\right), we henceforth set:

𝒀n=(Y1,Y2,,Yn)\bm{Y}_{\!\!n}=\big{(}Y_{1},Y_{2},\ldots,Y_{n}\big{)} (4)

In other words, 𝒀n\bm{Y}_{\!\!n} is the truncated version of the original sequence Y{Y} at the nnth term.

Definition 5 (Selectivity of a test).

Given any nn\in\mathbb{N} and any test f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right), the selectivity of ff at given level γ(0,1)\gamma\in(0,1) is defined as the set:

Selγ(f):={q[0,1/2):YSeqq,f(𝒀n)Decγ}\textup{Sel}_{\gamma}\left(f\right)\,\,:=\,\,\Big{\{}\,q\in[0,1/2):\forall\,{Y}\in\textup{Seq}_{\,q},f\left(\bm{Y}_{\!\!n}\right)\in\mathrm{Dec}_{\gamma}\,\Big{\}}

The relevance of the interval [0,1/2)[0,1/2) in the definition above will pop up in Section 6.2.

Definition 6 (Landscapes of tests).

Given any nn\in\mathbb{N} and any test f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right), the landscape of ff at given level γ(0,1)\gamma\in(0,1) is the subset of Decγ\,\mathrm{Dec}_{\gamma} defined by:

Lndγ(f):=_qSelγ(f){f(𝒀n):YSeqq}\mathrm{Lnd}_{\gamma}\left({f}\right)\,\,:=\,\,\bigcup_{\_}{q\in\textup{Sel}_{\gamma}\left(f\right)}\Big{\{}f\left(\bm{Y}_{\!\!n}\right):{Y}\in\textup{Seq}_{\,q}\Big{\}} (5)

The total landscape covered by all the tests f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) is defined by setting:

Lndscpγ:=_n{Lndγ(f):f(n,{0,1})}{\mathrm{Lndscp}_{\gamma}}\,\,\,\,:=\,\,\,\,\bigcup_{\_}{n\in\mathbb{N}}\,\Big{\{}\,\mathrm{Lnd}_{\gamma}\left({f}\right):f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right)\,\Big{\}} (6)

This notion of landscape makes it possible to compare tests via the following preorder. The proofs that the following definition is consistent and that the next lemma holds true are left to the reader.

Definition 7 (Preorder (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}}\,,\,\,{\preceq}^{\!*}\,\,)).

Given any level γ(0,1)\gamma\in(0,1), we define the preorder (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}}\,,\,\,{\preceq}^{\!*}\,\,) via the following three properties:
(P1) n\forall n\in\mathbb{N}, (f,g)(n,{0,1})×(n,{0,1})\forall(f,g)\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right)\times\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right), Lndγ(f)Lndγ(g)\mathrm{Lnd}_{\gamma}\left({f}\right)\,\,{\preceq}^{\!*}\,\,\mathrm{Lnd}_{\gamma}\left({g}\right) if:

Selγ(f)=Selγ(g)andqSelγ(f),YSeqq,f(𝒀n)g(𝒀n)\textup{Sel}_{\gamma}\left(f\right)=\textup{Sel}_{\gamma}\left(g\right)\quad\text{and}\quad\forall\,q\in\textup{Sel}_{\gamma}\left(f\right),\forall\,{Y}\in\textup{Seq}_{\,q},f\left(\bm{Y}_{\!\!n}\right)\,\,\,{\preceq}\,\,\,\,g\left(\bm{Y}_{\!\!n}\right)

(P2) (𝔏,𝔏)(Lndscpγ2𝕆γ)×2𝕆γ\forall\,(\mathfrak{L},\mathfrak{L}^{\prime})\in\left({\mathrm{Lndscp}_{\gamma}}\cup 2^{\mathbb{O}_{\gamma}}\right)\!\!\times 2^{\mathbb{O}_{\gamma}},  𝔏𝔏\mathfrak{L}\,\,{\preceq}^{\!*}\,\,\mathfrak{L}^{\prime}
(P3) 𝔏2Decγ(Lndscpγ2𝕆γ)\forall\,\mathfrak{L}\in 2^{{\mathrm{Dec}_{\gamma}}}\setminus\left({\mathrm{Lndscp}_{\gamma}}\cup 2^{\mathbb{O}_{\gamma}}\right), 𝔏𝔏\mathfrak{L}\,\,\,{\preceq}^{\!*}\,\,\,\mathfrak{L}

Lemma 2.

(𝔏,𝔏)(Lndscpγ2𝕆γ)×(Lndscpγ2𝕆γ)\forall\,(\mathfrak{L},\mathfrak{L}^{\prime})\in\left({\mathrm{Lndscp}_{\gamma}}\cup 2^{\mathbb{O}_{\gamma}}\right)\times\left({\mathrm{Lndscp}_{\gamma}}\cup 2^{\mathbb{O}_{\gamma}}\right), 𝔏𝔏𝔏×𝔏.\mathfrak{L}\,\,{\preceq}^{\!*}\,\,\mathfrak{L}^{\prime}\Rightarrow\mathfrak{L}\times\mathfrak{L}^{\prime}\subset\,\,{\preceq}\,\,.

With this material, we can state our first result that will prove useful in applications to statistical decisions below.

Theorem 1 (Approximation of oracles in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\!\,\,{\preceq}^{\!*}\,\,)).

Given γ(0,1)\gamma\in(0,1), if a set 𝔛γ\mathfrak{X}_{\gamma} and a family of tests (f_ξ,n)_ξ𝔛γ,n\left(f_{\_}{\xi,n}\right)_{\_}{\xi\in\mathfrak{X}_{\gamma},n\in\mathbb{N}} satisfy:
(i) (ξ,n)𝔛γ×\forall(\xi,n)\in\mathfrak{X}_{\gamma}\times\mathbb{N}, f_ξ,n(n,{0,1})f_{\_}{\xi,n}\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right);
(ii) 𝒬_γ[0,)\exists\mathcal{Q}_{\_}\gamma\subset[0,\infty), (ξ,n)𝔛γ×\forall(\xi,n)\in\mathfrak{X}_{\gamma}\times\mathbb{N}, Selγ(f_ξ,n)=𝒬_γ\textup{Sel}_{\gamma}\left(f_{\_}{\xi,n}\right)=\mathcal{Q}_{\_}\gamma;
(iii) (ξ,q)𝔛γ×𝒬_γ\forall(\xi,q)\in\mathfrak{X}_{\gamma}\times\mathcal{Q}_{\_}\gamma, YSeqq\forall{Y}\in\textup{Seq}_{\,q}, lim_ndet[f_ξ,n(𝐘n)]=1\displaystyle\lim_{\_}{n\to\infty}\mathbb{P}_{\textsc{det}}\left[{f_{\_}{\xi,n}\left(\bm{Y}_{\!\!n}\right)}\right]=1;
then, by setting Lndscpγ={Lndγ(f_ξ,n):n,ξ𝔛γ}{\mathrm{Lndscp}^{\prime}_{\gamma}}=\Big{\{}\,\mathrm{Lnd}_{\gamma}\left({f_{\_}{\xi,n}}\right):n\in\mathbb{N},\xi\in\mathfrak{X}_{\gamma}\,\Big{\}}, we have:

2𝕆γ=upper(Lndscpγ,(2Decγ,))=sup(Lndscpγ,(2Decγ,))2^{\mathbb{O}_{\gamma}}=\textup{upper}\left({\mathrm{Lndscp}^{\prime}_{\gamma}}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right)=\sup\left({\mathrm{Lndscp}^{\prime}_{\gamma}}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right) (7)
Proof.

For any (ξ,n)𝔛γ×(\xi,n)\in\mathfrak{X}_{\gamma}\times\mathbb{N} and any 𝔏2𝕆γ\mathfrak{L}\in 2^{\mathbb{O}_{\gamma}}, (P2) in Definition 7 straightforwardly implies that Lndγ(f_ξ,n)𝔏\mathrm{Lnd}_{\gamma}\left({f_{\_}{\xi,n}}\right)\,\,{\preceq}^{\!*}\,\,\mathfrak{L}. As a consequence:

2𝕆γupper(Lndscpγ,(2Decγ,))2^{\mathbb{O}_{\gamma}}\subset\textup{upper}\left({\mathrm{Lndscp}_{\gamma}}^{\prime}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right) (8)

To prove the converse inclusion, consider some 𝔏upper(Lndscpγ,(2Decγ,))\mathfrak{L}\in\textup{upper}\left({\mathrm{Lndscp}_{\gamma}}^{\prime}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right). We thus have (ξ,n)×𝔛γ\forall\,(\xi,n)\in\mathbb{N}\times\mathfrak{X}_{\gamma}, Lndγ(f_ξ,n)𝔏\mathrm{Lnd}_{\gamma}\left({f_{\_}{\xi,n}}\right)\,\,{\preceq}^{\!*}\,\,\mathfrak{L}. According to Lemma 2, we have (ξ,n)𝔛γ×\forall\,({\xi,n})\in\mathfrak{X}_{\gamma}\times\mathbb{N}, Lndγ(f_ξ,n)×𝔏\mathrm{Lnd}_{\gamma}\left({f_{\_}{\xi,n}}\right)\times\mathfrak{L}\subset\,\,{\preceq}\,\,. Therefore, (ξ,n)×𝔛γ\forall\,({\xi,n})\in\mathbb{N}\times\mathfrak{X}_{\gamma}, qSelγ(f_ξ,n)\forall\,q\in\textup{Sel}_{\gamma}\left(f_{\_}{\xi,n}\right), YSeqq\forall\,{Y}\in\textup{Seq}_{\,q} and D𝔏\forall\,{D}\in\mathfrak{L}, f_ξ,n(𝒀n)𝔏f_{\_}{\xi,n}\left(\bm{Y}_{\!\!n}\right)\,\,{\preceq}\,\,\mathfrak{L}. It follows from the definition of \,\,{\preceq}\,\, and assumption (ii) above that:

(ξ,n)×𝔛γ,q𝒬_γ,YSeqq,D𝔏,det[f_ξ,n(𝒀n)]det[D]\forall\,(\xi,n)\in\mathbb{N}\times\mathfrak{X}_{\gamma},\forall\,q\in\mathcal{Q}_{\_}\gamma,\forall\,{Y}\in\textup{Seq}_{\,q},\forall\,{D}\in\mathfrak{L},\mathbb{P}_{\textsc{det}}\left[{f_{\_}{\xi,n}\left(\bm{Y}_{\!\!n}\right)}\right]\leqslant\mathbb{P}_{\textsc{det}}\left[{{D}}\right]

We derive from assumption (iii) that det[D]=1\mathbb{P}_{\textsc{det}}\left[{{D}}\right]=1 and thus that D𝕆γ{D}\in\mathbb{O}_{\gamma}. It follows that 𝔏2𝕆γ\mathfrak{L}\in 2^{\mathbb{O}_{\gamma}}. We obtain that upper(Lndscpγ,(2Decγ,))2𝕆γ\textup{upper}\left({\mathrm{Lndscp}_{\gamma}}^{\prime}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right)\subset 2^{\mathbb{O}_{\gamma}} and therefore, from (8), 2𝕆γ=upper(Lndscpγ,(2Decγ,))2^{\mathbb{O}_{\gamma}}=\textup{upper}\left({\mathrm{Lndscp}_{\gamma}}^{\prime}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right). The second equality in (7) is straightforward since the elements of 2𝕆γ2^{\mathbb{O}_{\gamma}} are isomorphic in the sense of \,\,{\preceq}^{\!*}\,\,. ∎

For later use, given J[0,)J\subset[0,\infty), nn\in\mathbb{N} and 𝔉(n,{0,1})\mathfrak{F}\subset\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right), we hereafter set:

LndscpsγJ(𝔉):={Lndγ(f)Lndscpγ:f𝔉,Selγ(f)=J}\mathrm{Lndscps}^{\,J}_{\gamma}(\mathfrak{F})\,\,:=\,\,\Big{\{}\mathrm{Lnd}_{\gamma}\left({f}\right)\in{\mathrm{Lndscp}_{\gamma}}:f\in\mathfrak{F}\,,\,\textup{Sel}_{\gamma}\left(f\right)=J\Big{\}} (9)

5 The Neyman-Pearson (NP) solution

When nn spans \mathbb{N}, the Neyman-Pearson (NP) Lemma makes it possible to pinpoint a maximal element in each (Lndscpsγ{0}(𝔉),)(\mathrm{Lndscps}^{\{0\}}_{\gamma}(\mathfrak{F}),\,\,{\preceq}\,\,) with 𝔉=(n,{0,1})\mathfrak{F}=\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right). These maximal elements are hereafter called NP decisions. Specifically, we have the following result.

Lemma 3 (Maximality of the NP decisions).

For any γ(0,1)\gamma\in(0,1) and any nn\in\mathbb{N},

Lndγ(fnnp(γ))=max(Lndscpsγ{0}((n,{0,1})),)\mathrm{Lnd}_{\gamma}\left({f^{\textsc{np}(\gamma)}_{n}}\right)=\max\left(\mathrm{Lndscps}^{\{0\}}_{\gamma}\left(\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right)\right),\,\,{\preceq}^{\!*}\,\,\right) (10)

where fnnp(γ)(n,{0,1})f^{\textsc{np}(\gamma)}_{n}\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) is the nn-dimensional NP test with size γ\gamma defined by:

(y_1,y_2,,y_n)n,fnnp(γ)(y_1,y_2,,y_n)={1if _i=1ny_i>nΦ1(1γ)0otherwise\forall(y_{\_}1,y_{\_}2,\ldots,y_{\_}n)\in\mathbb{R}^{n},f^{\textsc{np}(\gamma)}_{n}(y_{\_}1,y_{\_}2,\ldots,y_{\_}n)=\left\{\begin{array}[]{lll}1&\text{if \, $\sum_{\_}{i=1}^{n}y_{\_}{i}>\sqrt{n}\,\Phi^{-1}(1-\gamma)$}\vskip 6.0pt plus 2.0pt minus 2.0pt\\ 0&\text{otherwise}\end{array}\right. (11)

and satisfies, YSeq 0\forall\,{Y}\in\textup{Seq}_{\,0},

{fa[fnnp(γ)(𝒀n)]=γdet[fnnp(γ)(𝒀n)]=1Φ(Φ1(1γ)n)\left\{\begin{array}[]{lll}\mathbb{P}_{\textsc{fa}}\left[{f^{\textsc{np}(\gamma)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]=\gamma\vskip 6.0pt plus 2.0pt minus 2.0pt\\ \mathbb{P}_{\textsc{det}}\left[{f^{\textsc{np}(\gamma)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]=1-\Phi\big{(}\,\Phi^{-1}(1-\gamma)-\sqrt{n}\,\big{)}\end{array}\right.
Proof.

A direct application of the Neyman-Pearson Lemma [9, Theorem 3.2.1, page 60], followed by some standard algebra to obtain det[fnnp(γ)(𝒀n)]\mathbb{P}_{\textsc{det}}\left[{f^{\textsc{np}(\gamma)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]. ∎

The next result states that it suffices to increase the number of observations to approximate oracles with level γ\gamma by NP decisions.

Theorem 2 (Approximation of oracles with level γ\gamma by NP decisions in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\!\,\,{\preceq}^{\!*}\,\,)).


Setting LndNP(γ):={Lndγ(fnnp(γ)):n}\,\mathrm{Lnd}^{\textup{NP}(\gamma)}\,\,:=\,\,\Big{\{}\,\mathrm{Lnd}_{\gamma}\left({f^{\textsc{np}(\gamma)}_{n}}\right):n\in\mathbb{N}\,\Big{\}} for any γ(0,1)\gamma\in(0,1), we have:

2𝕆γ=upper(LndNP(γ),(2Decγ,))=sup(LndNP(γ),(2Decγ,))2^{\mathbb{O}_{\gamma}}=\textup{upper}\left(\mathrm{Lnd}^{\textup{NP}(\gamma)}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right)=\sup\left(\mathrm{Lnd}^{\textup{NP}(\gamma)}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right)
Proof.

Given γ(0,1)\gamma\in(0,1), set 𝔛γ={0}\mathfrak{X}_{\gamma}=\{0\} and, n,f_0,n=fnnp(γ)\forall\,n\in\mathbb{N},f_{\_}{0,n}=f^{\textsc{np}(\gamma)}_{n}. According to Lemma 3:

lim_ndet[fnnp(γ)(𝒀n)]=1\displaystyle\lim_{\_}{n\to\infty}\mathbb{P}_{\textsc{det}}\left[{f^{\textsc{np}(\gamma)}_{n}(\bm{Y}_{\!\!n})}\right]=1

Thence the result as a consequence of Theorem 1. ∎

6 The RDT solution

6.1 An elementary RDT problem

Problem statement. The RDT theoretical framework is exposed in full details in [10, 11]. To ease the reading of the present paper, we directly focus on the particular RDT problem that can be used in connection with the detection problem at stake.

In this respect, suppose that Z=Θ+W(Ω,n)Z=\Theta+W\in\mathcal{M}(\Omega,\mathbb{R}^{n}), where Θ\Theta and WW are independent elements of (Ω,n)\mathcal{M}(\Omega,\mathbb{R}^{n}). In the sequel, we assume that W𝒩(0,𝐈n)W\thicksim\mathcal{N}(0,\mathbf{I}_{n}), 𝐈n\mathbf{I}_{n} being the n×nn\times n identity matrix, and consider the mean testing problem of deciding on whether |Θn(ω)|τ|\langle{\Theta}\rangle_{n}(\omega)|\leqslant\tau (null hypothesis _0\mathcal{H}_{\_}0) or |Θn(ω)|>τ|\langle{\Theta}\rangle_{n}(\omega)|>\tau (alternative hypothesis _1\mathcal{H}_{\_}1), when we are given Z(ω)=Θ(ω)+W(ω)Z(\omega)=\Theta(\omega)+W(\omega), for ωΩ\omega\in\Omega. The idea is that Θ\Theta oscillates uncontrollably around 0 and that only sufficient large deviations of the norm should be detected. This is a particular Block-RDT problem, following the terminology and definition given in [11]. This problem is summarized by dropping ω\omega, as usual, and writing:

{Observation:Z=Θ+W(Ω,n) with {Θ(Ω,n),W𝒩(0,𝐈_N),Θ and W are independent,_0:|Θn|τ,_1:|Θn|>τ.\left\{\!\!\!\!\begin{array}[]{lll}\,\text{{Observation:}}\,Z=\Theta+W\in\mathcal{M}(\Omega,\mathbb{R}^{n})\text{\, with }\left\{\!\!\!\begin{array}[]{lll}\Theta\in\mathcal{M}(\Omega,\mathbb{R}^{n}),W\thicksim\mathcal{N}(0,\mathbf{I}_{\_}N),\\ \text{$\Theta$ and $W$ are independent},\\ \end{array}\right.\\ \begin{array}[]{lll}\!\mathcal{H}_{\_}0:\,\,|\langle{\Theta}\rangle_{n}|\leqslant\tau,\\ \!\mathcal{H}_{\_}1:\,\,|\langle{\Theta}\rangle_{n}|>\tau.\end{array}\end{array}\right. (12)

Standard likelihood theory [9, 2, 3] does not make it possible to solve this problem. Fortunately, this problem can be solved as follows via the Random Distortion Testing (RDT) framework.
Size and power of tests for mean testing. We seek tests with guaranteed size and optimal power, in the sense specified below.

Definition 8 (Size for the mean testing problem).

The size of f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) for testing the empirical mean of the signals Θ(Ω,n)\Theta\in\mathcal{M}(\Omega,\mathbb{R}^{n}) such that [|Θn|τ]0\mathbb{P}\big{[}|\langle{\Theta}\rangle_{n}|\leqslant\tau\big{]}\neq 0, given Z=Θ+W(Ω,n)Z=\Theta+W\in\mathcal{M}(\Omega,\mathbb{R}^{n}) with WW independent of Θ\Theta, is defined by:

α[n](f)=sup_Θ(Ω,n):[|Θn|τ]0[f(Z)=1||Θn|τ]\alpha^{[n]}(f)=\sup_{\_}{\Theta\in\mathcal{M}(\Omega,\mathbb{R}^{n})\,:\,\mathbb{P}\,[\,|\langle{\Theta}\rangle_{n}|\leqslant\tau\,]\neq 0}\mathbb{P}\big{[}f(Z)=1\,\big{|}\,|\langle{\Theta}\rangle_{n}|\leqslant\tau\big{]} (13)

We say that f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) has level (resp. size) γ\gamma if α[n](f)γ\alpha^{[n]}(f)\leqslant\gamma (resp. α[n](f)=γ\alpha^{[n]}(f)=\gamma). The class of all the tests with level γ\gamma is denoted by Testsγ[n]\mathrm{Tests}^{[{n}]}_{\gamma}:

Testsγ[n]={f(n,{0,1}):α[n](f)γ}\mathrm{Tests}^{[{n}]}_{\gamma}=\left\{f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right):\alpha^{[n]}(f)\leqslant\gamma\right\}
Definition 9 (Power for the mean testing problem).

The power of f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) for testing the empirical mean of Θ(Ω,n)\,\,\Theta\in\mathcal{M}(\Omega,\mathbb{R}^{n}) such that [|Θn|>τ]0\mathbb{P}\big{[}|\langle{\Theta}\rangle_{n}|>\tau\big{]}\neq 0 when we are given Z=Θ+W(Ω,n)Z=\Theta+W\in\mathcal{M}(\Omega,\mathbb{R}^{n}), with WW independent of Θ\,\,\Theta, is defined by:

βΘ[n](f)=[f(Z)=1||Θn|>τ]\beta^{[n]}_{\Theta}(f)=\mathbb{P}\big{[}f(Z)=1\,\big{|}\,|\langle{\Theta}\rangle_{n}|>\tau\big{]} (14)

The RDT solution. With the same notation as above, we can easily construct a preorder (Testsγ[n],)\left(\mathrm{Tests}^{[{n}]}_{\gamma},\,\,{\preceq}^{\diamond}\,\,\right) by setting:

(f,f)Testsγ[n]×Testsγ[n],\displaystyle\forall\,(f,f^{\prime})\in\mathrm{Tests}^{[{n}]}_{\gamma}\times\mathrm{Tests}^{[{n}]}_{\gamma},
ffifΘ\displaystyle f\,\,{\preceq}^{\diamond}\,\,f^{\prime}\,\,\text{if}\,\,\,\forall\,\,\,\Theta (Ω,n),[|Θn|>τ]0βΘ[n](f)βΘ[n](f)\displaystyle\in\mathcal{M}(\Omega,\mathbb{R}^{n}),\mathbb{P}\big{[}|\langle{\Theta}\rangle_{n}|>\tau\big{]}\neq 0\Rightarrow\beta^{[n]}_{\Theta}(f)\leqslant\beta^{[n]}_{\Theta}(f^{\prime})

No maximal element exists in (Testsγ[n],)\left(\mathrm{Tests}^{[{n}]}_{\gamma},\,\,{\preceq}^{\diamond}\,\,\right). However, we can exhibit γ[n]Testsγ[n]\mathfrak{C}^{[n]}_{\gamma}\subset\mathrm{Tests}^{[{n}]}_{\gamma} whose elements satisfy suitable invariance properties with respect to the mean testing problem and prove the existence of a maximal element in (γ[n],)\left(\mathfrak{C}^{[n]}_{\gamma},\,\,{\preceq}^{\diamond}\,\,\right).

Set 𝒮={id,id}\mathcal{S}=\big{\{}\textrm{id},-\textrm{id}\big{\}} where id is the identity of \mathbb{R}. Endowed with the usual composition law \circ of functions, (𝒮,)(\mathcal{S},\circ) is a group. Let 𝒜\mathcal{A} be the group action that associates to each given s𝒮{s}\in\mathcal{S} the map 𝒜_s:nn\mathcal{A}_{\_}{s}:\mathbb{R}^{n}\rightarrow\mathbb{R}^{n} defined for every x=(x_1,x_2,,x_n)nx=(x_{\_}1,x_{\_}2,\ldots,x_{\_}n)\in\mathbb{R}^{n} by 𝒜_s(x)=(s(x_1),s(x_2),,s(x_n)).\mathcal{A}_{\_}{s}(x)=({s}(x_{\_}1),{s}(x_{\_}2),\ldots,{s}(x_{\_}n)). Readily, the mean testing problem is invariant under the action of 𝒜\mathcal{A} in that 𝒜_s(Z)=𝒜_s(Θ)+W\mathcal{A}_{\_}{s}(Z)=\mathcal{A}_{\_}{s}(\Theta)+W^{\prime} where W=(W_1,W_2,,W_n)𝒩(0,𝐈_n)W^{\prime}=(W^{\prime}_{\_}1,W^{\prime}_{\_}2,\ldots,W^{\prime}_{\_}n)\thicksim\mathcal{N}(0,\mathbf{I}_{\_}n) is independent of 𝒜_s(Θ)\mathcal{A}_{\_}{s}(\Theta). Therefore, 𝒜_s(Z)\mathcal{A}_{\_}{s}(Z) satisfies the same hypotheses as ZZ. We also have |𝒜_s(Θ)n|=|Θn||\langle{\mathcal{A}_{\_}{s}(\Theta)}\rangle_{n}|=|\langle{\Theta}\rangle_{n}|. Hence, the mean testing problem remains unchanged by substituting 𝒜_s(Θ)\mathcal{A}_{\_}{s}(\Theta) for Θ\Theta and WW^{\prime} for WW. It is thus natural to seek 𝒜\mathcal{A}-invariant tests, that is, tests f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) such that f(𝒜_s(x))=f(x)f(\mathcal{A}_{\_}{s}(x))=f(x) for any s𝒮{s}\in\mathcal{S} and any xnx\in\mathbb{R}^{n}.

On the other hand, since we can reduce the noise variance by averaging observations, we consider 𝒜\mathcal{A}-invariant integrator tests, that is, 𝒜\mathcal{A}-invariant tests f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) for which exists f¯(1,{0,1})\overline{f}\in\mathcal{M}\left(\mathbb{R}^{1},\{0,1\}\right), henceforth called the reduced form of ff, such that f(𝒙)=f¯(𝒙n)f(\bm{x})=\overline{f}(\langle{\bm{x}}\rangle_{n}) for any 𝒙n\bm{x}\in\mathbb{R}^{n}. Reduced forms of 𝒜\mathcal{A}-invariant integrator tests are also 𝒜\mathcal{A}-invariant: x\forall x\in\mathbb{R}, s𝒜\forall{s}\in\mathcal{A}, f¯(s(x))=f¯(x)\overline{f}({s}(x))=\overline{f}(x). We thus define γ[n]Testsγ[n]\mathfrak{C}^{[n]}_{\gamma}\subset\mathrm{Tests}^{[{n}]}_{\gamma} as the class of all 𝒜\mathcal{A}-invariant integrator tests with level γ\gamma. We thus have fγ[n]f\in\mathfrak{C}^{[n]}_{\gamma} if:
[Size]: α[n](f)γ\alpha^{[n]}(f)\leqslant\gamma;
[𝒜\mathcal{A}-invariance]: (s,x)𝒮×n\forall\,({s},x)\in\mathcal{S}\times\mathbb{R}^{n}, f(𝒜_s(x))=f(x)f\left(\mathcal{A}_{\_}{s}(x)\right)=f(x);
[Integration]: f¯(1,{0,1})\exists\,\overline{f}\in\mathcal{M}\left(\mathbb{R}^{1},\{0,1\}\right), xn\forall\,x\in\mathbb{R}^{n}, f(x)=f¯(xn)f(x)=\overline{f}(\langle{x}\rangle_{n}).
The following result derives from the foregoing and [10, 11].

Proposition 2 (Maximal element of (γ[n],)\left(\mathfrak{C}^{[n]}_{\gamma},\,\,{\preceq}^{\diamond}\,\,\right)).

For any γ(0,1)\gamma\in(0,1) and any nn\in\mathbb{N},

{fnrdt(γ,τ)}=max(γ[n],)\left\{f^{\textsc{rdt}(\gamma,\tau)}_{n}\right\}=\max\left(\mathfrak{C}^{[n]}_{\gamma},\,\,{\preceq}^{\diamond}\,\,\right) (15)

where fnrdt(γ,τ)(n,{0,1})f^{\textsc{rdt}(\gamma,\tau)}_{n}\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right) is defined by setting

(y_1,y_2,,y_n)n,fnrdt(γ,τ)(y_1,y_2,,y_n)={1if |_i=1ny_i|nλγ(τn)0otherwise\forall(y_{\_}1,y_{\_}2,\ldots,y_{\_}n)\in\mathbb{R}^{n},f^{\textsc{rdt}(\gamma,\tau)}_{n}(y_{\_}1,y_{\_}2,\ldots,y_{\_}n)=\left\{\begin{array}[]{lll}1&\text{if \, \, $\left|\sum_{\_}{i=1}^{n}y_{\_}i\right|\leqslant\sqrt{n}\,\lambda_{\gamma}(\tau\sqrt{n})$}\vskip 6.0pt plus 2.0pt minus 2.0pt\\ 0&\text{otherwise}\end{array}\right.

and λγ(τn)\lambda_{\gamma}(\tau\sqrt{n}) is the unique solution in xx to the equation

2Φ(xτn)Φ(x+τn)=γ2-\Phi(x-\tau\sqrt{n})-\Phi(x+\tau\sqrt{n})=\gamma

where Φ\Phi is the cumulative distribution function (cdf) of the 𝒩(0,1)\mathcal{N}(0,1) law.

RDT and NP tests are structurally different because dedicated to two different testing problems and optimal with respect to two different criteria. This structural difference will be enhanced by coming back to our initial detection problem.

6.2 Application to Detection

Consider again the problem of estimating ε{0,1}\varepsilon\in\{0,1\}, when we have a sequence YSeqq{Y}\in\textup{Seq}_{\,q} of observations such that:

n,(ε,ω){0,1}×Ω,Yn(ε,ω)=ε+Δ_n(ω)+X_n(ω)\forall\,n\in\mathbb{N},\forall(\varepsilon,\omega)\in\big{\{}0,1\big{\}}\times\Omega,Y_{n}(\varepsilon,\omega)=\varepsilon+\Delta_{\_}n(\omega)+X_{\_}n(\omega) (16)

where X_1,X_2,iid𝒩(0,1)X_{\_}1,X_{\_}2,\ldots\stackrel{{\scriptstyle\text{iid}}}{{\thicksim}}\mathcal{N}(0,1) and n\forall n\in\mathbb{N}, Δ_n(q)\Delta_{\_}n\in\mathcal{B}_{\infty}(q) with q[0,)q\in[0,\infty). The empirical mean of YY satisfies: n,Yn(ε)=Y(ε)n=ε+Δn+Xn\forall\,n\in\mathbb{N},\langle{Y\,}\rangle_{n}(\varepsilon)=\langle{Y(\varepsilon)}\rangle_{n}=\varepsilon+\langle{{\Delta}}\rangle_{n}+\langle{X}\rangle_{n}. We thus have |Δn|q|\langle{{\Delta}}\rangle_{n}|\leqslant q (a-s). Set Θ_n=ε+Δ_n\Theta_{\_}n=\varepsilon+\Delta_{\_}n for every nn\in\mathbb{N}. In the sequel, we assume q<1/2q<1/2 because, in this case, we straightforwardly verify that

{ε=0|Θn|qε=1|Θn|1q\left\{\begin{array}[]{lll}\varepsilon=0&\Leftrightarrow&|\langle{\Theta}\rangle_{n}|\leqslant q\\ \varepsilon=1&\Leftrightarrow&|\langle{\Theta}\rangle_{n}|\geqslant 1-q\\ \end{array}\right. (17)

Therefore, when q[0,1/2)q\in[0,1/2), deciding on whether ε\varepsilon is zero or not when we are given 𝒀n(ω)\bm{Y}_{\!\!n}(\omega) amounts to testing whether |Θn(ω)|τ|\langle{\Theta}\rangle_{n}(\omega)|\leqslant\tau or not for τ[q,1q]\tau\in[q,1-q]. We thus can use the decision fnrdt(γ,τ)(𝒀n)f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right), where fnrdt(γ,τ)f^{\textsc{rdt}(\gamma,\tau)}_{n} is given by Proposition 2.

We can calculate the false alarm probability (1) of fnrdt(γ,τ)(𝒀n)f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right) where 𝒀n\bm{Y}_{\!\!n} is defined by (4). The theoretical results in [10] yield that τ[q,1q],fa[fnrdt(γ,τ)(𝒀n)]γ.\forall\tau\in[q,1-q],\mathbb{P}_{\textsc{fa}}\left[{f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]\leqslant\gamma. In the sequel, for the sake of simplifying notation, we assume that both τ\tau and qq are in [0,1/2)[0,1/2). In this case, we have:

τ[0,1/2),{Selγ(fnrdt(γ,τ))=[0,τ]Lndγ(fnrdt(γ,τ))=_q[0,τ]{fnrdt(γ,τ)(𝒀n):YSeqq}\forall\tau\in[0,1/2)\,,\,\,\left\{\begin{array}[]{lll}\textup{Sel}_{\gamma}\left(f^{\textsc{rdt}(\gamma,\tau)}_{n}\right)=[0,\tau]\vskip 6.0pt plus 2.0pt minus 2.0pt\\ \mathrm{Lnd}_{\gamma}\left({f^{\textsc{rdt}(\gamma,\tau)}_{n}}\right)=\displaystyle\bigcup_{\_}{q\in[0,\tau]}\Big{\{}f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right):{Y}\in\textup{Seq}_{\,q}\Big{\}}\end{array}\right. (18)

We can then state the following lemma, which is the counterpart to Lemma 3.

Theorem 3 (Maximality of RDT decisions).

For any γ(0,1)\gamma\in(0,1), any nn\in\mathbb{N} and any 0qτ<1/20\leqslant q\leqslant\tau<1/2, Lndγ(fnrdt(γ,τ))=max(Lndscpsγ[0,τ](γ[n]),)\mathrm{Lnd}_{\gamma}\left({f^{\textsc{rdt}(\gamma,\tau)}_{n}}\right)=\max\left(\mathrm{Lndscps}^{[0,\tau]}_{\gamma}\left(\mathfrak{C}^{[n]}_{\gamma}\right),\,\,{\preceq}^{\!*}\,\,\right).

Proof.

It results from Definition 6 that Lndγ(f):={f(𝒀n):YSeqq,q[0,τ]}\mathrm{Lnd}_{\gamma}\left({f}\right)\,\,:=\,\,\Big{\{}f\left(\bm{Y}_{\!\!n}\right):{Y}\in\textup{Seq}_{\,q},q\in[0,\tau]\Big{\}}. According to (9), we also have:

Lndscpsγ[0,τ](γ[n])={Lndγ(f)Lndscpγ:fγ[n],Selγ(f)=[0,τ]}\mathrm{Lndscps}^{[0,\tau]}_{\gamma}\left(\mathfrak{C}^{[n]}_{\gamma}\right)=\Big{\{}\mathrm{Lnd}_{\gamma}\left({f}\right)\in{\mathrm{Lndscp}_{\gamma}}:f\in\mathfrak{C}^{[n]}_{\gamma}\,,\,\textup{Sel}_{\gamma}\left(f\right)=[0,\tau]\Big{\}}

Given q[0,τ]q\in[0,\tau] and YSeqq{Y}\in\textup{Seq}_{\,q}, set:

{Z=𝒀n=(Y_1,Y_2,,Y_n)(see (4))W=(X_1,X_2,,X_n)𝒩(0,𝐈n)Θ=(1+Δ_1,1+Δ_2,,1+Δ_n)\left\{\begin{array}[]{lllll}Z&=&\bm{Y}_{\!\!n}=\left({Y}_{\_}1,{Y}_{\_}2,\ldots,{Y}_{\_}n\right)\ \text{(see \eqref{Eq: vec of seq})}\\ W&=&(X_{\_}1,X_{\_}2,\ldots,X_{\_}n)\thicksim\mathcal{N}(0,\mathbf{I}_{n})\\ \Theta&=&(1+\Delta_{\_}1,1+\Delta_{\_}2,\ldots,1+\Delta_{\_}n)\end{array}\right.

We basically have Z=Θ+WZ=\Theta+W. Consider now the mean testing problem (12) with Θ\Theta, WW and ZZ defined as above. For any f(n,{0,1})f\in\mathcal{M}\left(\mathbb{R}^{n},\{0,1\}\right), it follows from Eqs. (16) , (17), (2) and (13) that:

βΘ[n](f)=det[f(𝒀n)]\beta^{[n]}_{\Theta}(f)=\mathbb{P}_{\textsc{det}}\left[{f\left(\bm{Y}_{\!\!n}\right)}\right] (19)

Suppose now that fγ[n]f\in\mathfrak{C}^{[n]}_{\gamma} with Selγ(f)=[0,τ]\textup{Sel}_{\gamma}\left(f\right)=[0,\tau]. We derive from Proposition 2, (19) and its application to fnrdt(γ,τ)f^{\textsc{rdt}(\gamma,\tau)}_{n}, that det[f(𝒀n)]det[fnrdt(γ,τ)(𝒀n)].\mathbb{P}_{\textsc{det}}\left[{f\left(\bm{Y}_{\!\!n}\right)}\right]\leqslant\mathbb{P}_{\textsc{det}}\left[{f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]. Since qτ<1/2q\leqslant\tau<1/2 implies that qSelγ(f)q\in\textup{Sel}_{\gamma}\left(f\right) and since Selγ(f)=Selγ(fnrdt(γ,τ))=[0,τ]\textup{Sel}_{\gamma}\left(f\right)=\textup{Sel}_{\gamma}\left(f^{\textsc{rdt}(\gamma,\tau)}_{n}\right)=[0,\tau], we can rewrite the foregoing equality as f(𝒀n)fnrdt(γ,τ)(𝒀n)f\left(\bm{Y}_{\!\!n}\right)\,\,{\preceq}\,\,f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right). This holding true for any qSelγ(f)q\in\textup{Sel}_{\gamma}\left(f\right), any YSeqq{Y}\in\textup{Seq}_{\,q} and since ff and fqrdt(γ,n)f^{\textsc{rdt}(\gamma,n)}_{q} have same selectivity [0,τ][0,\tau], we derive from the foregoing and Definition 7 that Lndγ(f)Lndγ(fnrdt(γ,τ)).\mathrm{Lnd}_{\gamma}\left({f}\right)\,\,{\preceq}^{\!*}\,\,\mathrm{Lnd}_{\gamma}\left({f^{\textsc{rdt}(\gamma,\tau)}_{n}}\right).

We now prove that the oracles with level γ\gamma are approximated by RDT decisions.

Lemma 4 (Approximation of oracles with γ\gamma by RDT decisions in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,)).


Setting LndRDT(γ,τ):={Lndγ(fnrdt(γ,τ)):n}\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\,\,:=\,\,\Big{\{}\,\mathrm{Lnd}_{\gamma}\left({f^{\textsc{rdt}(\gamma,\tau)}_{n}}\right):n\in\mathbb{N}\,\Big{\}} for any given γ(0,1)\gamma\in(0,1), we have:

2𝕆γ=upper(LndRDT(γ,τ),(2Decγ,))=sup(LndRDT(γ,τ),(2Decγ,))2^{\mathbb{O}_{\gamma}}=\textup{upper}\left(\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right)=\sup\left(\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,\big{)}\right)
Proof.

Given γ(0,1)\gamma\in(0,1), it follows from (2) and [10, Theorem 2] that:

(q,τ)[0,1/2)×[0,1/2),n,det[fnrdt(γ,τ)(𝒀n)]Q_1/2((1q)n,λγ(τn))\begin{array}[]{ccc}\forall(q,\tau)\in[0,1/2)\times[0,1/2),\forall n\in\mathbb{N},\mathbb{P}_{\textsc{det}}\left[{f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]\geqslant Q_{\_}{1/2}\big{(}(1-q)\sqrt{n},\lambda_{\gamma}(\tau\sqrt{n})\big{)}\end{array}

Since τ<1q\tau<1-q, [8, Eq. (3) and Lemma B.2] induce that lim_ndet[fnrdt(γ,τ)(𝒀n)]=1\displaystyle\lim_{\_}{n\to\infty}\mathbb{P}_{\textsc{det}}\left[{f^{\textsc{rdt}(\gamma,\tau)}_{n}\left(\bm{Y}_{\!\!n}\right)}\right]=1. The set LndRDT(γ,τ)Lndscpγ\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\subset{\mathrm{Lndscp}_{\gamma}} thus satisfies Theorem 1 conditions with 𝔛γ={τ}\mathfrak{X}_{\gamma}=\{\tau\} and n,f_n,τ=fnrdt(γ,τ)\forall\,n\in\mathbb{N},f_{\_}{n,\tau}=f^{\textsc{rdt}(\gamma,\tau)}_{n}. ∎

7 Multiplicity Principle in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,)

To state the MP in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,), we need the following lemma.

Lemma 5 (Selectivity of NP tests).

n\forall n\in\mathbb{N}, Selγ(fnnp(γ))={0}\textup{Sel}_{\gamma}\left(f^{\textsc{np}(\gamma)}_{n}\right)=\{0\}

Proof.

A consequence of [10, Section B, p. 6.]. ∎

We have now all the material to state the main result.

Theorem 4 (Multiplicity Principle in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,)).

For any given τ(0,1/2)\tau\in(0,1/2), the MP is satisfied in (2Decγ,)(2^{{\mathrm{Dec}_{\gamma}}},\,\,{\preceq}^{\!*}\,\,) by the pair (LndNP(γ),LndRDT(γ,τ))\left(\mathrm{Lnd}^{\textup{NP}(\gamma)},\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\right).

Proof.

According to Theorems 2 and 3, the subsets LndNP(γ)\mathrm{Lnd}^{\textup{NP}(\gamma)} and LndRDT(γ,τ)\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)} of 2Decγ2^{{\mathrm{Dec}_{\gamma}}} are such that

sup(LndNP(γ),(2Decγ,))=sup(LndRDT(γ,τ),(2Decγ,))=2𝕆γ\sup\left(\mathrm{Lnd}^{\textup{NP}(\gamma)}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}}\,,\,\,{\preceq}^{\!*}\,\,\big{)}\right)=\sup\left(\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\,\,,\,\big{(}2^{{\mathrm{Dec}_{\gamma}}}\,,\,\,{\preceq}^{\!*}\,\,\big{)}\right)=2^{\mathbb{O}_{\gamma}}

In addition, (18) and Lemma 5 imply that LndNP(γ)×LndRDT(γ,τ)=\mathrm{Lnd}^{\textup{NP}(\gamma)}\times\mathrm{Lnd}^{\textup{RDT}(\gamma,\tau)}\cap\,\,{\preceq}^{\!*}\,\,=\emptyset. The conclusion follows from Lemma 1. ∎

8 Conclusions and Perspectives

In this paper, via the framework provided by the Multiple Principle (MP), which is motivated by the concept of degeneracy in biology, and by introducing the notions of test landscapes and selectivity, we have established that this principle is satisfied when we consider the standard NP tests and the RDT tests applied to a detection problem. One interest of this result is that it opens prospects on the construction of Memory Evolutive Systems [5, 6] via tests.

More elaborated statistical decision problems should be considered beyond this preliminary work. Sequential tests are particularly appealing because they collect information till they can decide with guaranteed performance bounds. On the one hand, the Sequential Probability Ratio Test (SPRT) established in [12] is proved to be optimal; on the other hand, in [8], we have exhibited non-optimal tests with performance guarantees in presence of interferences. In the same way as NP and RDT tests satisfy PM, we conjecture that these two types of sequential tests satisfy MP as well.

From a pratical point of view, such results open new prospects for the design of networks of sensors, where combining different types of sensors and tests satisfying the MP could bring resilience to the overall system.

Acknowledgements

The authors are very grateful to the reviewers for their strong encouragements and insightful remarks that help improve the readiness of this paper.

References