This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

How to compute the volume in low dimension?

Arjan Cornelissen Université Paris Cité, CNRS, IRIF, Paris, France Simons Institute, UC Berkeley, California, USA Simon Apers Université Paris Cité, CNRS, IRIF, Paris, France Sander Gribling Tilburg University, Tilburg, the Netherlands
Abstract

Estimating the volume of a convex body is a canonical problem in theoretical computer science. Its study has led to major advances in randomized algorithms, Markov chain theory, and computational geometry. In particular, determining the query complexity of volume estimation to a membership oracle has been a longstanding open question. Most of the previous work focuses on the high-dimensional limit. In this work, we tightly characterize the deterministic, randomized and quantum query complexity of this problem in the high-precision limit, i.e., when the dimension is constant.

1 Introduction

The volume estimation problem is a widely-studied problem, that lies on the interface between convex geometry and computational complexity. The typical setting is the following, see for instance [Sim03]. We consider a convex body KdK\subseteq\mathbb{R}^{d} that satisfies BdKRBdB_{d}\subseteq K\subseteq RB_{d}, where Bd={xd:x1}B_{d}=\{x\in\mathbb{R}^{d}:\left\|x\right\|\leq 1\} denotes the unit ball in dd dimensions, and RBdRB_{d} the radius-RR ball in dd dimensions. Denoting the volume of KK by Vol(K)\operatorname{Vol}(K), the aim of volume estimation is to output V~0\widetilde{V}\geq 0 such that |V~Vol(K)|<εVol(K)|\widetilde{V}-\operatorname{Vol}(K)|<\varepsilon\operatorname{Vol}(K), for a given relative precision ε>0\varepsilon>0.

In this work, we are interested in the query complexity of volume estimation. Specifically, we will assume access to KK through a membership oracle which, after querying a point xdx\in\mathbb{R}^{d}, returns whether xKx\in K. The key question is then how many membership queries (as a function of dd, RR and ε\varepsilon) are required to solve the volume estimation problem. We analyze this question in three computational models, namely the deterministic, randomized and quantum models, each of which is more powerful than the last.

Previous work mostly studied query complexity in the high-dimensional setting, which tracks the dependency of the query complexity on the dimension dd. The title of this work is a reference to the early survey by Simonovits called “How to compute the volume in high dimension?” [Sim03], and we refer the interested reader to this survey for a more in-depth discussion of this setting.

In this work, however, we focus on the low-dimensional limit, or high-precision limit, i.e., the limit where the dimension dd is fixed and the precision ε\varepsilon goes to zero. While we are not aware of any literature targeting this regime in the membership oracle setting, there are numerous algorithms in different access models targeting this regime. We summarize the most relevant of these below, and refer the interested reader to the excellent survey by Brunel [Bru18].

Low-dimensional computational geometry.

Before turning to volume estimation, we discuss a related problem that has been well studied in the low-dimensional limit. In the convex set estimation problem, one tries to output a convex set K~d\widetilde{K}\subseteq\mathbb{R}^{d} that estimates the original convex set KK.

One way to measure the error in this problem is through the relative Nikodym metric, defined as the volume of the symmetric difference K~ΔK\widetilde{K}\Delta K divided by the volume of KK. An interesting sequence of results in this setting starts with a paper by Schütt [Sch94]. They considered a bounded convex set KRBddK\subseteq RB_{d}\subseteq\mathbb{R}^{d} with RO(1)R\in O(1), and showed that the convex hull of nn points sampled uniformly at random from KK, approximates it up to relative Nikodym distance O(n2/(d+1))O(n^{-2/(d+1)}). Later, Brunel proved a matching lower bound [Bru16], implying that in order to achieve relative error ε\varepsilon, one requires Θ(ε(d+1)/2)\Theta(\varepsilon^{-(d+1)/2}) samples from the uniform distribution over KK to solve the convex set estimation problem. Surprisingly, Baldin and Reiß [BR16] showed that the volume estimation problem is significantly easier in this setting, i.e., they show that Θ(ε2(d+1)/(d+3))\Theta(\varepsilon^{-2(d+1)/(d+3)}) uniform samples from KK are necessary and sufficient to estimate its volume up to relative error ε\varepsilon.

Another line of research considers a qualitatively stronger error metric for the convex set estimation problem. To that end, suppose that a convex body K~d\widetilde{K}\subseteq\mathbb{R}^{d} is an approximation of a convex body KdK\subseteq\mathbb{R}^{d}. For any ε(0,1)\varepsilon\in(0,1), Agarwal, Har-Peled and Varadarajan [AHPV04] define K~\widetilde{K} to be an ε\varepsilon-kernel of KK if K~K\widetilde{K}\subseteq K, and

uSd1,maxxK~uTxminxK~uTx(1ε)[maxxKuTxminxKuTx],\forall u\in S_{d-1},\qquad\max_{x\in\widetilde{K}}u^{T}x-\min_{x\in\widetilde{K}}u^{T}x\geq(1-\varepsilon)\left[\max_{x\in K}u^{T}x-\min_{x\in K}u^{T}x\right],

where Sd1=BdS_{d-1}=\partial B_{d} is the unit sphere in dd dimensions. If K~\widetilde{K} is an ε\varepsilon-kernel of KK, then the Hausdorff distance between K~\widetilde{K} and KK, i.e., the maximum distance between a point in KK from K~\widetilde{K} and vice versa, is at most εdiam(K)\varepsilon\operatorname{diam}(K) (see Lemma 2.7). Being an ε\varepsilon-kernel is invariant under taking arbitrary linear transformations, just like the Nikodym metric, and any ε\varepsilon-kernel of KK approximates it up to relative Nikodym distance O(ε)O(\varepsilon) too (see Proposition 2.8), yet the converse does not hold.

The starting point towards algorithmically constructing such ε\varepsilon-kernels is an early paper by Dudley [Dud74, HJ19], which proves that every convex body KBdK\subseteq B_{d} has an approximate polytope K~K\widetilde{K}\subseteq K with Hausdorff distance at most ε\varepsilon and only O(ε(d1)/2)O(\varepsilon^{-(d-1)/2}) faces. Independently, Brohnsteyn and Ivanov [BI75] proved that a similar approximate polytope exists with only O(ε(d1)/2)O(\varepsilon^{-(d-1)/2}) vertices. Subsequently, in the more restricted setting where KK is well-rounded, i.e., BdKRBdB_{d}\subseteq K\subseteq RB_{d}, Agarwal, Har-Peled and Varadarajan [AHPV04] used Brohnsteyn and Ivanov’s construction to produce an ε\varepsilon-kernel with O((R/ε)(d1)/2)O((R/\varepsilon)^{(d-1)/2}) vertices, and their algorithm was subsequently improved independently by [Cha06] and [YAPV08]. The referenced resources only explicitly analyze their algorithms’ performance in terms of the runtime in the case where KK is the convex hull of nn points, with [YAPV08] ultimately achieving a runtime of O~(n+ε1/2)\widetilde{O}(n+\varepsilon^{-1/2}) for d=2d=2 and O~(n+εd2)\widetilde{O}(n+\varepsilon^{d-2}) for d3d\geq 3. In this work, we build on these algorithmic ideas and port them to the membership oracle setting.

Our results.

In this paper, we tightly characterize the membership oracle query complexity for the convex set estimation and volume estimation problems, in the low-dimensional limit, and in the deterministic, randomized and quantum models. The resulting complexities are displayed in Table 1.1.

Problem Convex set estimation Volume estimation
Error metric Constructing ε\varepsilon-kernel Rel. Nikodym distance ε\varepsilon Rel. error ε\varepsilon
Deterministic Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}}) Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}}) Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}})
Randomized Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}}) Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}}) Θ~(ε2(d1)d+3)\widetilde{\Theta}(\varepsilon^{-\frac{2(d-1)}{d+3}})
Quantum Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}}) Θ~(εd12)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{2}}) Θ~(εd1d+1)\widetilde{\Theta}(\varepsilon^{-\frac{d-1}{d+1}})
Table 1.1: Overview of the query complexity results obtained in this paper for the convex set estimation and volume estimation problems. We only track the dependence on 1/ε1/\varepsilon and RR, the prefactors are allowed to depend exponentially on the dimension dd. The tilde hides polylogarithmic factors in 1/ε1/\varepsilon and RR.

For the convex set estimation problem, we consider both the problem of constructing an ε\varepsilon-kernel, as well as estimating KK up to relative Nikodym distance ε\varepsilon. We obtain that the query complexities of both problems are Θ~(ε(d1)/2)\widetilde{\Theta}(\varepsilon^{-(d-1)/2}) in all three computational models. This shows in particular that randomness and quantumness alike do not provide any benefit over the deterministic model. Furthermore, for estimating a convex set up to relative Nikodym error, having access to a membership oracle is strictly more powerful than uniform sampling from the convex body, as we beat the Θ(ε(d+1)/2)\Theta(\varepsilon^{-(d+1)/2}) samples that are required in that setting [Bru16].

For the volume estimation problem, we plot the obtained complexities in Figure 1.1. For any fixed dimension dd, we beat the previously best-known algorithms in the randomized and quantum settings, that respectively make O(1/ε2)O(1/\varepsilon^{2}) and O(1/ε)O(1/\varepsilon) queries. The exponents in our complexities converge to these naive bounds as dd increases, showing that the obtained advantage becomes ever smaller with increasing dd. Moreover, the gap between the randomized and quantum query complexity is small when dd is small, and becomes bigger as dd increases, converging to a full quadratic separation in the regime where dd is large. Like in the uniform sampling setting, the volume estimation problem is significantly easier than the convex set estimation problem in the randomized model.

Θ~(ε)\widetilde{\Theta}(\varepsilon^{-\cdot})ddDeterministic, \rightarrow\inftyRandomized, 2\rightarrow 2Quantum, 1\rightarrow 122334455667788991010334455Quantum, 1Randomized, 2Previous boundsNew results
Figure 1.1: Graph of the exponents in the query complexities for the volume estimation problem. dd is fixed, and the asymptotic limit is for RR\to\infty and ε0\varepsilon\downarrow 0. The tilde hides polylogarithmic factors in 1/ε1/\varepsilon and RR. The dashed lines represent the previously best-known results, and the solid ones connect the newly-found complexities.

Techniques.

Our algorithms are based on the approaches taken in [AHPV04, Cha06, YAPV08], and they all proceed in a similar manner. First, we apply an existing, deterministic rounding procedure [GLS88, Theorem 4.6.1] that makes O~(1)\widetilde{O}(1) queries and transforms the convex body into a “well-rounded” one, i.e., satisfying BdKRBdB_{d}\subseteq K\subseteq RB_{d} with RO(1)R\in O(1) when dd is fixed.

Then, we follow [YAPV08], and take a set {vj}j=1n\{v_{j}\}_{j=1}^{n} of nn roughly equally spaced points on the boundary of (R+1)Bd(R+1)B_{d}, with nΘ(ε(d1)/2)n\in\Theta(\varepsilon^{-(d-1)/2}). Every point vjv_{j} is subsequently projected onto the convex body, and the convex hull of the resulting points forms an ε\varepsilon-kernel of KK, as was shown in [AHPV04]. Since these projections cannot be implemented directly in our model, we formulate every projection operation as a convex optimization problem, and use existing solvers from [GLS88] to obtain an approximate projection. It was already shown in [Cha06] and [YAPV08, Theorem 1] that this suffices to obtain an ε\varepsilon-kernel K~\widetilde{K} of KK. It also follows that K~\widetilde{K} is an ε\varepsilon-precise approximation of KK in the relative Nikodym metric.

Next, we find a speed-up of this approach for the volume estimation problem. To that end, we first run the convex set estimation procedure to some precision ε>ε\varepsilon^{\prime}>\varepsilon, to obtain an ε\varepsilon^{\prime}-kernel K¯\underline{K} of KK. We slightly enlarge the inner approximation K¯K\underline{K}\subseteq K to generate an outer approximation K¯K\overline{K}\supseteq K of KK, with the property that Vol(K¯K¯)/Vol(K)O(ε)\operatorname{Vol}(\overline{K}\setminus\underline{K})/\operatorname{Vol}(K)\in O(\varepsilon^{\prime}). Then, in the randomized setting, we sample Θ((ε/ε)2)\Theta((\varepsilon^{\prime}/\varepsilon)^{2}) times uniformly from the region K¯K¯\overline{K}\setminus\underline{K}, and compute the fraction of samples ξ\xi that is inside KK. We then use this as a refinement to our estimate of Vol(K)\operatorname{Vol}(K), i.e., we output Vol(K¯)+ξVol(K¯K¯)\operatorname{Vol}(\underline{K})+\xi\operatorname{Vol}(\overline{K}\setminus\underline{K}). In the quantum case, we employ the same idea, but this time we use quantum amplitude estimation to quadratically speed up the computation of ξ\xi. In both cases, balancing the number of queries made in the first and second step yields the complexity bounds from Table 1.1.

The matching lower bounds for the query complexities follow from a construction that dates back to Dudley [Dud74], where it was used to prove a lower bound on the number of facets needed for approximating convex bodies by polytopes. The construction packs nn disjoint, identical spherical caps on the boundary of the unit ball through a δ\delta-net, for some suitable value of δ\delta that depends on dd and ε\varepsilon. Then, a convex body KxBdK_{x}\subseteq B_{d} is associated to a bitstring xx of length nn by including the jjth spherical cap in KxK_{x} if and only if xj=1x_{j}=1. See Figure 4.1 for an illustration.

The core observation is that estimating KxK_{x}, or its volume, becomes equivalent to learning a fraction of the entries of the bitstring xx, or its Hamming weight, respectively. Additionally, a membership query to KxK_{x} can be simulated with a single query to xx. Recovering a constant fraction of bits of an nn-bit string is known to cost Θ(n)\Theta(n) queries in all three models that we consider. Similarly, the problem of estimating the Hamming weight of a bitstring is also known as the approximate counting problem and its (deterministic, randomized and quantum) query complexity has been tightly characterized in earlier work. We show that by carefully choosing the size of the spherical caps, this construction provides matching lower bounds for the query complexities from Table 1.1 up to polylogarithmic factors in RR and ε\varepsilon (and large factors in the dimension dd).

History in the high-dimensional setting.

For completeness, we also comment on volume estimation in the high-dimensional setting, which tracks the dependency of the query complexity on the dimension dd. The title of this work is a reference to the early survey by Simonovits called “How to compute the volume in high dimension?” [Sim03], and we refer the interested reader to this survey for a more in-depth discussion of this setting. The main takeaways in this setting are the following:

  • (i)

    In the deterministic setting, any deterministic algorithm must make a number of queries that is exponential in dd [Ele86, BF87].

  • (ii)

    In the randomized setting, there famously exists a Markov chain Monte Carlo algorithm that makes a number of queries polynomially in dd and 1/ε1/\varepsilon [DFK91]. While the initial bound scaled as O~(d23/ε2)\widetilde{O}(d^{23}/\varepsilon^{2}), a long of research led to the current best bound that is O~(d3/ε2)\widetilde{O}(d^{3}/\varepsilon^{2}). This bound follows from combining algorithms in [CV18, JLLV21] with recent breakthroughs on the so-called KLS-constant (see e.g. [KL22]), and it approaches a lower bound of Ω~(d2)\widetilde{\Omega}(d^{2}) on the randomized query complexity [RV08].

  • (iii)

    In the quantum setting, Chakrabarti, Childs, Hung, Li, Wang and Wu [CCH+23] gave an improved bound of the form O~(d3+d2.5/ε)\widetilde{O}(d^{3}+d^{2.5}/\varepsilon).111In [CCH+23, Page 20:11], the authors claim that recent breakthroughs in the KLS-conjecture [Che21, JLV22, Kla23] improve the analysis of their volume estimation algorithm to O~(d2.5+o(1)+d2+o(1)/ε)\widetilde{O}(d^{2.5+o(1)}+d^{2+o(1)}/\varepsilon) membership oracle queries. They argue that the mixing time of the hit-and-run walk is O~(d2ψ2)\widetilde{O}(d^{2}\psi^{2}), where ψ\psi is the KLS-constant. This indeed appears to be a correct upper bound on the mixing time of the ball walk, see [JLLV21, Theorem 2.7], but, as far as we are aware, the mixing time of the hit-and-run walk is not known to be improved due to the resolution of the KLS-conjecture. Adapting the algorithm of [CCH+23] to make use of the ball walk is highly non-trivial. Consequently, the question of whether volume estimation can be done in O~(d2.5+o(1)+d2+o(1)/ε)\widetilde{O}(d^{2.5+o(1)}+d^{2+o(1)}/\varepsilon) quantum queries to the membership oracle is still open. This was later improved to O~(d3+d2.25/ε)\widetilde{O}(d^{3}+d^{2.25}/\varepsilon) by Cornelissen and Hamoudi [CH23]. Both use quantum algorithms to speed up the Markov chain Monte Carlo method. The authors of [CCH+23] also prove that the query complexity is Ω(d)\Omega(\sqrt{d}) when ε\varepsilon is constant, and Ω(1/ε)\Omega(1/\varepsilon) in the regime where 1/dε1/31/d\leq\varepsilon\leq 1/3, which in turn implies Ω(d)\Omega(d) when ε=1/d\varepsilon=1/d.222In [CCH+23], the authors additionally informally suggest a lower bound of Ω(1/ε)\Omega(1/\varepsilon) in general, but the corresponding theorem statement excludes the limit ε0\varepsilon\downarrow 0. A lower bound of Ω(1/ε)\Omega(1/\varepsilon) for fixed dd and in the limit where ε0\varepsilon\downarrow 0 would indeed contradict our results.

Organization.

In Section 2, we fix notation, formally define algorithmic models and computational problems, and state results from (computational) geometry. Subsequently, in Section 3 we develop the algorithms, and in Section 4 we prove the corresponding lower bounds.

2 Preliminaries

2.1 Notation

We start by fixing some notation. Let \mathbb{N} be the set of all positive integers. For all nn\in\mathbb{N}, let [n]={1,,n}[n]=\{1,\dots,n\}, and [n]0={0}[n][n]_{0}=\{0\}\cup[n]. For any kk\in\mathbb{N} (or xx\in\mathbb{R}), we let k\mathbb{N}_{\geq k} (resp. x\mathbb{R}_{\geq x}) be the set of all integers (resp. reals) that are bigger than or equal to kk (resp. xx). We denote the Euclidean norm by \left\|\cdot\right\|. For every dd\in\mathbb{N}, we denote the unit ball in dd dimensions by Bd={xd:x1}B_{d}=\{x\in\mathbb{R}^{d}:\left\|x\right\|\leq 1\}. We write Γd=Vol(Bd)\Gamma_{d}=\operatorname{Vol}(B_{d}) for the volume of BdB_{d}. We denote the unit sphere in dd dimensions by Sd1=BdS_{d-1}=\partial B_{d}. For two sets AA and BB we let AΔB=(AB)(BA)A\Delta B=(A\setminus B)\cup(B\setminus A) be the symmetric difference of AA and BB.

A set KdK\subseteq\mathbb{R}^{d} is convex if for any two points x,yKx,y\in K, the straight line segment between xx and yy is fully contained in KK, i.e., for every λ[0,1]\lambda\in[0,1], λx+(1λ)yK\lambda x+(1-\lambda)y\in K. For two convex sets K,KdK,K^{\prime}\subseteq\mathbb{R}^{d}, we recall that KKK\cap K^{\prime} and K+K={x+y:xK,yK}K+K^{\prime}=\{x+y:x\in K,y\in K^{\prime}\} are again convex. For ease of notation, we typically make an implicit assumption that all convex bodies are closed.

We use big-OO-notation to hide constant factors. That is, for two functions f,g:0d0f,g:\mathbb{R}_{\geq 0}^{d}\to\mathbb{R}_{\geq 0}, we write fO(g)f\in O(g) if there exist C,r>0C,r>0 such that f(x)Cg(x)f(x)\leq Cg(x) whenever xr\left\|x\right\|\geq r. We write fΩ(g)f\in\Omega(g) if gO(f)g\in O(f), and we write fΘ(g)f\in\Theta(g) if fO(g)Ω(g)f\in O(g)\cup\Omega(g). Sometimes, the limit in which the big-OO-notation is to be interpreted plays an important role. For instance if the big-OO-notation holds in the limit where ε0\varepsilon\downarrow 0, then we interpret the function f,gf,g as functions in the variable 1/ε1/\varepsilon.

Similarly, we use the big-OO-tilde-notation to hide polylogarithmic factors. That is, if f,g:0d0f,g:\mathbb{R}_{\geq 0}^{d}\to\mathbb{R}_{\geq 0}, then we write fO~(g)f\in\widetilde{O}(g) if there exist constants C,k,r0C,k,r\geq 0 such that f(x)Cg(x)j=1dlogk(xj)f(x)\leq Cg(x)\prod_{j=1}^{d}\log^{k}(x_{j}) whenever xr\left\|x\right\|\geq r. We write fΩ~(g)f\in\widetilde{\Omega}(g) if gO~(f)g\in\widetilde{O}(f), and fΘ~(g)f\in\widetilde{\Theta}(g) if fO~(g)Ω~(g)f\in\widetilde{O}(g)\cup\widetilde{\Omega}(g).

2.2 Access model and computational models

The input to the algorithms we design in this work are a dimension dd\in\mathbb{N}, outer radius R1R\geq 1, precision ε(0,1)\varepsilon\in(0,1), and a convex body KdK\subseteq\mathbb{R}^{d}. The access model specifies how the algorithm can access this input. Throughout, we will assume that the parameters dd, RR and ε\varepsilon are set beforehand, and hence are known to the algorithm. As for the convex body KK, many ways to access it have been considered in the literature. We refer to [GLS88, Chapter 2] for an overview of several different access models. In this work, we assume that it can be accessed by means of a (strong) membership oracle, in line with [GLS88, Definition 2.1.5].

Definition 2.1 (Membership oracle).

Let dd\in\mathbb{N}, and KdK\subseteq\mathbb{R}^{d} be a convex body. A membership oracle to KK is a procedure that, given a point xdx\in\mathbb{R}^{d}, outputs whether xKx\in K.

In general, an oracle is a subroutine that the algorithm can run only as a black box, i.e., it has no more information about what happens when it is run. One such call to the oracle is referred to as a query. The query complexity of a problem is the minimum number of queries any algorithm needs to make to solve said problem.

Additional to the access model, we also specify computational models that define the operations that the algorithms are allowed to use. We consider three different computational models in this work. In the deterministic model, the algorithm is completely described beforehand and follows a deterministic computational path. In the randomized model, the algorithm is allowed to use randomness during the execution of the algorithm, and in particular let the oracle’s inputs depend on it. It is required to output a correct answer with probability at least 2/32/3. In the quantum model, the algorithm is additionally allowed to perform operations on a quantum state, and supply the oracle’s inputs in a superposition, which it answers coherently. For a more detailed introduction to quantum algorithms, we refer to [NC10].

Throughout the execution of our algorithms, we additionally assume that we can store vectors in d\mathbb{R}^{d} precisely in memory in all three models, and that we can perform basic arithmetic with them exactly. We also assume in the randomized model that we can sample from any distribution on d\mathbb{R}^{d} as long as we have a classical description of it, and we use this to sample uniformly from an arbitrary region in Theorem 3.4. Similarly, in the quantum model, we assume that we can generate a uniform superposition over an arbitrary region in d\mathbb{R}^{d}, and we use this in Theorem 3.5. Both of these operations do not depend on KK, so in an implementation of the algorithms presented in this work, one can suppress the finite-precision errors that arise from these assumptions arbitrarily without increasing the number of membership oracle queries.

2.3 Rounding convex bodies

We start by describing a routine that “rounds” the convex body. A similar idea appears in [AHPV04, Cha06, YAPV08], but in contrast to their approach, we use inner and outer ellipses, rather than inner and outer cubes in our rounding procedure. This difference is not fundamental, it merely allows us to use a rounding routine that already exists in the literature [GLS88].

The aim of “rounding” a convex body KdK\subseteq\mathbb{R}^{d} is to find an invertible affine linear map L:xx0+TxL:x\mapsto x_{0}+Tx, such that BdL(K)RBdB_{d}\subseteq L(K)\subseteq R^{\prime}B_{d} with R1R^{\prime}\geq 1 as small as possible. Intuitively, one can think of rounding as finding a linear transformation that compresses the convex body as much as possible, so that it does not have parts that stick out far from the origin. Most of the existing literature focuses on randomized algorithms to compute this affine linear map [LV06, JLLV21], with the current state-of-the-art obtaining RO~(d)R^{\prime}\in\widetilde{O}(\sqrt{d}) and requiring O~(d3)\widetilde{O}(d^{3}) queries. In our setting, though, we require a deterministic procedure, and since we take dd to be a constant we don’t necessarily need RR^{\prime} to scale optimally in the dimension. Below, we sketch how to obtain a deterministic rounding procedure that obtains RO(d3)R^{\prime}\in O(d^{3}) and runs in time O(poly(d,log(R)))O(\operatorname{poly}(d,\log(R))).

Theorem 2.2 ([GLS88]).

Let dd\in\mathbb{N}, R1R\geq 1, and let KdK\subseteq\mathbb{R}^{d} be convex such that BdKRBdB_{d}\subseteq K\subseteq RB_{d}. Then, there is a deterministic algorithm that makes O(poly(d,log(R))O(\operatorname{poly}(d,\log(R)) queries, and finds an invertible affine linear map LL such that BdL(K)RBdB_{d}\subseteq L(K)\subseteq R^{\prime}B_{d}, with R=d(d+1)2O(d3)R^{\prime}=d(d+1)^{2}\in O(d^{3}). When dd is fixed, the algorithm makes O~(1)\widetilde{O}(1) queries, and RO(1)R^{\prime}\in O(1).

Proof.

Since KdK\subseteq\mathbb{R}^{d} satisfies BdKRBdB_{d}\subseteq K\subseteq RB_{d}, we observe from [GLS88, Figure 4.1] that we can deterministically turn a membership oracle into a weak separation oracle. The total multiplicative factor incurred from this conversion is O(poly(d,log(R)))O(\operatorname{poly}(d,\log(R))), and hence is O~(1)\widetilde{O}(1) in the case where dd is fixed.

Subsequently, from [GLS88, Theorem 4.6.1], we find that with O(poly(d,log(R)))O(\operatorname{poly}(d,\log(R))) calls to a weak separation oracle to KK, we can find an ellipsoid EE, such that EKd(d+1)2EE\subseteq K\subseteq d(d+1)^{2}E. Thus, by letting LL be the affine linear transformation that maps EE to the unit ball, we obtain that

BdL(K)RBd,whereR=d(d+1)2O(d3).B_{d}\subseteq L(K)\subseteq R^{\prime}B_{d},\qquad\text{where}\qquad R^{\prime}=d(d+1)^{2}\in O(d^{3}).

For fixed dd, we obtain RO(1)R^{\prime}\in O(1), and the query complexity of finding LL is in O~(1)\widetilde{O}(1). ∎

2.4 Geometry

We proceed by recalling some theoretical background in geometry. We start by formally introducing the concept of a net.

Definition 2.3 (δ\delta-net).

Let dd\in\mathbb{N}, δ>0\delta>0, and let SdS\subseteq\mathbb{R}^{d} be any set. We say that NSN\subseteq S is a δ\delta-net of SS if

  1. 1.

    For any two distinct points v,wNv,w\in N, we have vwδ\left\|v-w\right\|\geq\delta.

  2. 2.

    For any vSv\in S, there exists a wNw\in N such that vwδ\left\|v-w\right\|\leq\delta.

Intuitively, one can think of a δ\delta-net NN as the centers of a set of balls of radius δ\delta, that together cover the set SS. Moreover, since these centers are at least separated by distance δ\delta, the balls with radius δ/2\delta/2 centered at the points in NN must be disjoint up to a measure-zero set. By comparing the surface area of the sphere to that of {uSd1:uvδ}\{u\in S_{d-1}:\|u-v\|\leq\delta\} (for some fixed vSd1v\in S_{d-1}), the following bound follows on the number of points in a δ\delta-net on the sphere. We stress that the assumption that dd is fixed is crucial in the above proposition. Indeed, the Θ\Theta-notation hides a prefactor that might depend exponentially on dd.

Proposition 2.4.

Let dd\in\mathbb{N} be fixed. Let δ(0,1)\delta\in(0,1), and let NδN_{\delta} be a δ\delta-net of Sd1S_{d-1}. Then, |Nδ|Θ(δ(d1))|N_{\delta}|\in\Theta(\delta^{-(d-1)}).

Proof.

For any vVv\in V and r>0r>0, let Rv,r={wSd1:vwr}R_{v,r}=\{w\in S_{d-1}:\left\|v-w\right\|\leq r\}. The surface areas of Rv,rR_{v,r} is Av,rΘ(rd1)A_{v,r}\in\Theta(r^{d-1}) and the surface area of Sd1S_{d-1} is Ad1Θ(1)A_{d-1}\in\Theta(1). Since NδN_{\delta} is a δ\delta-net, all Rv,δ/2R_{v,\delta/2}’s with vNδv\in N_{\delta} are disjoint up to possibly some measure-zero set, and all Rv,δR_{v,\delta}’s cover Sd1S_{d-1}. Thus, we find |Nδ|Av,δ/2Ad1|Nδ|Av,δ|N_{\delta}|A_{v,\delta/2}\leq A_{d-1}\leq|N_{\delta}|A_{v,\delta}, from which we conclude that |Nδ|Θ(δ(d1))|N_{\delta}|\in\Theta(\delta^{-(d-1)}). ∎

We proceed with formally defining an ε\varepsilon-kernel.

Definition 2.5 (ε\varepsilon-kernel [AHPV04]).

Let dd\in\mathbb{N}, and ε(0,1)\varepsilon\in(0,1). Let KdK\subseteq\mathbb{R}^{d} be a convex body. Then a convex body K~K\widetilde{K}\subseteq K is an ε\varepsilon-kernel of KK if

uSd1,maxxK~uTxminxK~uTx(1ε)[maxxKuTxminxKuTx].\forall u\in S_{d-1},\qquad\max_{x\in\widetilde{K}}u^{T}x-\min_{x\in\widetilde{K}}u^{T}x\geq(1-\varepsilon)\left[\max_{x\in K}u^{T}x-\min_{x\in K}u^{T}x\right].

One very desirable property of ε\varepsilon-kernels is its invariance under affine linear transformations.

Lemma 2.6 ([AHPV04],[AHP24, Lemma 2.5]).

Let dd\in\mathbb{N}, ε(0,1)\varepsilon\in(0,1), and let LL be an invertible affine linear map on d\mathbb{R}^{d}. Let K~,Kd\widetilde{K},K\subseteq\mathbb{R}^{d} be convex sets. Then, K~\widetilde{K} is an ε\varepsilon-kernel of KK if and only if L(K~)L(\widetilde{K}) is an ε\varepsilon-kernel of L(K)L(K).

Next, we make a connection between ε\varepsilon-kernels and approximations w.r.t. the Hausdorff metric.

Lemma 2.7.

Let dd\in\mathbb{N}, R1R\geq 1, ε>0\varepsilon>0, and let K,K~dK,\widetilde{K}\subseteq\mathbb{R}^{d} be convex sets.

  1. 1.

    If BdKB_{d}\subseteq K and K~K\widetilde{K}\subseteq K is an ε\varepsilon-precise Hausdorff approximation of KK, then K~\widetilde{K} is also an ε\varepsilon-kernel of KK.

  2. 2.

    If KRBdK\subseteq RB_{d} and K~\widetilde{K} is an ε\varepsilon-kernel of KK, it is a 2εR2\varepsilon R-precise Hausdorff approximation of KK.

Proof.

Observe that if K~K\widetilde{K}\subseteq K, then the Hausdorff distance dd between K~\widetilde{K} and KK is

d=maxuSd1[maxxKuTxmaxxK~uTx]=maxuSd1[minxK~uTxminxKuTx].d=\max_{u\in S_{d-1}}\left[\max_{x\in K}u^{T}x-\max_{x\in\widetilde{K}}u^{T}x\right]=\max_{u\in S_{d-1}}\left[\min_{x\in\widetilde{K}}u^{T}x-\min_{x\in K}u^{T}x\right]. (1)

For the first claim, we take uSd1u\in S_{d-1} arbitrarily, and observe from Eq. 1 that

maxxK~uTxminxK~uTxmaxxKuTxminxKuTx2ε(1ε)[maxxKuTxminxKuTx].\max_{x\in\widetilde{K}}u^{T}x-\min_{x\in\widetilde{K}}u^{T}x\geq\max_{x\in K}u^{T}x-\min_{x\in K}u^{T}x-2\varepsilon\geq(1-\varepsilon)\left[\max_{x\in K}u^{T}x-\min_{x\in K}u^{T}x\right].

For the second claim, let dd be the Hausdorff distance between KK and K~\widetilde{K}, and uu the vector that maximizes the middle expression in Eq. 1. Then, we have

d\displaystyle d =maxxKuTxmaxxK~uTx[maxxKuTxminxKuTx][maxxK~uTxminxK~uTx]\displaystyle=\max_{x\in K}u^{T}x-\max_{x\in\widetilde{K}}u^{T}x\leq\left[\max_{x\in K}u^{T}x-\min_{x\in K}u^{T}x\right]-\left[\max_{x\in\widetilde{K}}u^{T}x-\min_{x\in\widetilde{K}}u^{T}x\right]
ε[maxxKuTxminxKuTx]εdiam(K)2εR.\displaystyle\leq\varepsilon\left[\max_{x\in K}u^{T}x-\min_{x\in K}u^{T}x\right]\leq\varepsilon\cdot\operatorname{diam}(K)\leq 2\varepsilon R.\qed

Finally, we observe that an ε\varepsilon-kernel naturally provides a relative Nikodym approximation of KK, and can be used to construct an outer approximation of KK, in the following proposition.

Proposition 2.8.

Let dd\in\mathbb{N}, ε(0,1/(4d(d+1)2))\varepsilon\in(0,1/(4d(d+1)^{2})), and let K~\widetilde{K} be an ε\varepsilon-kernel of a convex set KdK\subseteq\mathbb{R}^{d}. Then, given a full description of K~\widetilde{K}, we can construct a convex set K¯K\overline{K}\supseteq K such that Vol(K¯K~)/Vol(K)(1+4εd(d+1)2)d1O(ε)\operatorname{Vol}(\overline{K}\setminus\widetilde{K})/\operatorname{Vol}(K)\leq(1+4\varepsilon d(d+1)^{2})^{d}-1\in O(\varepsilon), with O~(1)\widetilde{O}(1) membership oracles to KK. In particular, this implies that Vol(K~ΔK)/Vol(K)O(ε)\operatorname{Vol}(\widetilde{K}\Delta K)/\operatorname{Vol}(K)\in O(\varepsilon).

Proof.

Since being an ε\varepsilon-kernel and being a relative Nikodym-distance approximation are both invariant under affine linear transformations, we can without loss of generality first round the convex body KK using Theorem 2.2, using just O~(1)\widetilde{O}(1) membership queries. It then suffices to consider the case where BdKRBdB_{d}\subseteq K\subseteq RB_{d}, with R=d(d+1)2O(1)R=d(d+1)^{2}\in O(1). Since ε<1/(4R)\varepsilon<1/(4R), we find by Lemma 2.7 that BdKK~+2εRBdK~+Bd/2B_{d}\subseteq K\subseteq\widetilde{K}+2\varepsilon RB_{d}\subseteq\widetilde{K}+B_{d}/2, and so Bd/2K~B_{d}/2\subseteq\widetilde{K}. Thus,

Bd/2K~KK~+2εRBd(1+4εR)K~=:K¯,B_{d}/2\subseteq\widetilde{K}\subseteq K\subseteq\widetilde{K}+2\varepsilon RB_{d}\subseteq(1+4\varepsilon R)\widetilde{K}=:\overline{K},

from which we find that

Vol(K¯K~)Vol(K)=Vol(K¯)Vol(K~)Vol(K)=[(1+4εR)d1]Vol(K~)Vol(K)(1+4εR)d1O(ε).\frac{\operatorname{Vol}(\overline{K}\setminus\widetilde{K})}{\operatorname{Vol}(K)}=\frac{\operatorname{Vol}(\overline{K})-\operatorname{Vol}(\widetilde{K})}{\operatorname{Vol}(K)}=\frac{[(1+4\varepsilon R)^{d}-1]\operatorname{Vol}(\widetilde{K})}{\operatorname{Vol}(K)}\leq(1+4\varepsilon R)^{d}-1\in O(\varepsilon).\qed

2.5 Problem definitions

We consider three computational problems in this paper. We formally introduce them here.

Definition 2.9 (Problem definitions).

Let dd\in\mathbb{N}, R>1R>1, and ε(0,1)\varepsilon\in(0,1). Let KdK\subseteq\mathbb{R}^{d} be a convex body such that BdKRBdB_{d}\subseteq K\subseteq RB_{d}. We consider three problems:

  1. 1.

    The ε\varepsilon-kernel construction problem is the problem of outputting an ε\varepsilon-kernel of KK.

  2. 2.

    The ε\varepsilon-Nikodym construction problem is the problem of outputting a convex body K~d\widetilde{K}\subseteq\mathbb{R}^{d} such that Vol(KΔK~)εVol(K)\operatorname{Vol}(K\Delta\widetilde{K})\leq\varepsilon\operatorname{Vol}(K).

  3. 3.

    The volume estimation problem is the problem of outputting a non-negative real V~0\widetilde{V}\geq 0 such that |Vol(K)V~|εVol(K)|\operatorname{Vol}(K)-\widetilde{V}|\leq\varepsilon\operatorname{Vol}(K).

It follows directly that these problems are qualitatively decreasing in terms of their query complexity. That is, solving the ε\varepsilon-kernel construction problem also solves the Nikodym construction problem with precision O(ε)O(\varepsilon), by virtue of Proposition 2.8, which in turn solves the volume estimation problem with the same precision, since we can simply output the volume of the approximation K~\widetilde{K}. These relations are less clear if one considers the runtime instead, e.g., since computing the volume might be very costly.

3 Algorithms

3.1 Convex set estimation

In this subsection, we develop a deterministic algorithm that constructs an ε\varepsilon-kernel of a well-rounded convex set KdK\subseteq\mathbb{R}^{d}, using membership queries to it. The algorithm follows the same general strategy as in [YAPV08], and we replace their approximate nearest-neighbor queries by a query-efficient approximate projection onto a convex body. For this, we use the well-known observation that projection onto a convex body can be phrased as a convex optimization problem, which we can deterministically solve approximately with only polylogarithmically many membership oracles calls, using for example the ellipsoid method [GLS88].

Proposition 3.1.

Let dd\in\mathbb{N}, R1R\geq 1, ε>0\varepsilon>0, KdK\subseteq\mathbb{R}^{d} convex such that BdKRBdB_{d}\subseteq K\subseteq RB_{d}. Let xdKx\in\mathbb{R}^{d}\setminus K, and yy^{\prime} its projection onto KK. There is a deterministic algorithm that obtains an element y~K\widetilde{y}\in K such that y~yε\left\|\widetilde{y}-y^{\prime}\right\|\leq\varepsilon. The algorithm makes O(poly(d,log(x),log(R),log(1/ε)))O(\operatorname{poly}(d,\log(\left\|x\right\|),\log(R),\log(1/\varepsilon))) queries to a membership oracle of KK. When dd is fixed and RO(1)R\in O(1), this complexity is O~(1)\widetilde{O}(1).

Now, we are ready to state the ε\varepsilon-kernel construction algorithm. We start by considering the well-rounded case, in Algorithm 1. We subsequently prove its properties in Theorem 3.2.

Algorithm 1 Well-rounded deterministic ε\varepsilon-kernel construction [YAPV08]

Input:

  1. 1.

    d2d\in\mathbb{N}_{\geq 2}: the dimension.

  2. 2.

    ε(0,1)\varepsilon\in(0,1): the desired precision.

  3. 3.

    R1R\geq 1: the outer radius, with RO(1)R\in O(1).

  4. 4.

    OKO_{K}: a membership oracle that on input xdx\in\mathbb{R}^{d} returns whether xKx\in K, where BdKRBdB_{d}\subseteq K\subseteq RB_{d}.

Derived constant: η=ε/R\eta=\sqrt{\varepsilon/R}.

Output: An ε\varepsilon-kernel K~\widetilde{K} of KK.

Number of queries: O~(ε(d1)/2)\widetilde{O}(\varepsilon^{-(d-1)/2})\quad (in the limit where dd is fixed).

Procedure: RoundedEpsKernConstr(d,ε,R,OK)\texttt{RoundedEpsKernConstr}(d,\varepsilon,R,O_{K}):

  1. 1.

    Let {xj}j=1n\{x_{j}\}_{j=1}^{n} be an η\eta-net of (R+1)Sd1(R+1)S_{d-1}.

  2. 2.

    For j=1,,nj=1,\dots,n,

    1. (a)

      Project xjx_{j} onto KK with precision ε\varepsilon, using Proposition 3.1. Denote the outcome by pjp_{j}.

  3. 3.

    Output conv({pj}j=1n)\operatorname{conv}(\{p_{j}\}_{j=1}^{n}).

Theorem 3.2.

Let d2d\in\mathbb{N}_{\geq 2}, R1R\geq 1 with RO(1)R\in O(1), and KdK\subseteq\mathbb{R}^{d} be a convex body such that BdKRBdB_{d}\subseteq K\subseteq RB_{d}. Then, Algorithm 1 computes an ε\varepsilon-kernel of KK that satisfies K~K\widetilde{K}\subseteq K, with O~(ε(d1)/2)\widetilde{O}(\varepsilon^{-(d-1)/2}) membership oracle queries.

Proof.

Correctness is proven in [YAPV08, Theorem 1]. For the bound on the number of queries, observe from Proposition 2.4 that the η\eta-net contains O(η(d1))=O(ε(d1)/2)O(\eta^{-(d-1)})=O(\varepsilon^{-(d-1)/2}) points. Furthermore, for every point, we perform one approximate projection, which costs O~(1)\widetilde{O}(1) queries by Proposition 3.1. As such, we conclude that the total number of membership oracle queries is O~(ε(d1)/2)\widetilde{O}(\varepsilon^{-(d-1)/2}). ∎

In the case where KK is not well-rounded, we combine Algorithm 1 with the rounding procedure of Theorem 2.2, and conversion to Nikodym distance in Proposition 2.8, to obtain the following corollary.

Corollary 3.3 (Deterministic ε\varepsilon-kernel construction).

Let d2d\in\mathbb{N}_{\geq 2}, R1R\geq 1, ε(0,1/(4d(d+1)2))\varepsilon\in(0,1/(4d(d+1)^{2})), and KdK\subseteq\mathbb{R}^{d} be convex such that BdKRBdB_{d}\subseteq K\subseteq RB_{d}. Then, we can deterministically compute an ε\varepsilon-kernel K~\widetilde{K} of KK with O~(ε(d1)/2)\widetilde{O}(\varepsilon^{-(d-1)/2}) membership oracle queries. K~\widetilde{K} is also an O(ε)O(\varepsilon)-precise approximation of KK in the relative Nikodym distance, and Vol(K~)\operatorname{Vol}(\widetilde{K}) is an O(ε)O(\varepsilon)-precise relative estimate of Vol(K)\operatorname{Vol}(K).

This corollary proves all the O~(ε(d1)/2)\widetilde{O}(\varepsilon^{-(d-1)/2}) upper bounds in Table 1.1. In Section 4.2, we prove that this approach is essentially optimal for the reconstruction problem in all computational models, i.e., even in the randomized and quantum settings we cannot improve significantly over the approach taken in Corollary 3.3.

3.2 Volume estimation

In this subsection, we switch to the volume estimation problem. To start off, we remark that Corollary 3.3 solves it in the deterministic setting. In the randomized and quantum settings, however, we can obtain a significant improvement over this approach.

The core idea is to run the (deterministic) ε\varepsilon^{\prime}-kernel construction algorithm up to some worse precision ε>ε\varepsilon^{\prime}>\varepsilon, to obtain an inner approximation K¯K\underline{K}\subseteq K. We can the use Proposition 2.8 to obtain an outer approximation K¯K\overline{K}\supseteq K, such that Vol(K¯K¯)/Vol(K)O(ε)\operatorname{Vol}(\overline{K}\setminus\underline{K})/\operatorname{Vol}(K)\in O(\varepsilon^{\prime}).

Subsequently, in the randomized setting, we sample uniformly from K¯K¯\overline{K}\setminus\underline{K}, and compute the fraction of points ξ\xi that is inside KK. We then compute Vol(K¯)+ξVol(K¯K¯)\operatorname{Vol}(\underline{K})+\xi\operatorname{Vol}(\overline{K}\setminus\underline{K}) as a refined estimate of the volume of KK, eventually resulting in the following theorem.

Theorem 3.4 (Randomized volume estimation).

Let d2d\in\mathbb{N}_{\geq 2}, R1R\geq 1, ε(0,1)\varepsilon\in(0,1), KdK\subseteq\mathbb{R}^{d} convex such that and BdKRBdB_{d}\subseteq K\subseteq RB_{d}. We can compute V~0\widetilde{V}\geq 0 such that |V~Vol(K)|εVol(K)|\widetilde{V}-\operatorname{Vol}(K)|\leq\varepsilon\operatorname{Vol}(K) with probability at least 2/32/3, using a randomized algorithm that makes O~(ε2(d1)/(d+3))\widetilde{O}(\varepsilon^{-2(d-1)/(d+3)}) queries to a membership oracle of KK.

Proof.

We first use Corollary 3.3 to obtain an ε\varepsilon^{\prime}-kernel K¯K\underline{K}\subseteq K with

ε=(1+ε4d+3)1d14d(d+1)2Θ(ε4d+3).\varepsilon^{\prime}=\frac{(1+\varepsilon^{\frac{4}{d+3}})^{\frac{1}{d}}-1}{4d(d+1)^{2}}\in\Theta\left(\varepsilon^{\frac{4}{d+3}}\right).

We make O~((ε)(d1)/2)=O~(ε2(d1)/(d+3))\widetilde{O}((\varepsilon^{\prime})^{-(d-1)/2})=\widetilde{O}(\varepsilon^{-2(d-1)/(d+3)}) membership oracles in this step of the algorithm. Then, we use Proposition 2.8 to generate an outer approximation K¯K\overline{K}\supseteq K, using just O~(1)\widetilde{O}(1) membership oracle queries. We find that Vol(K¯/K¯)/Vol(K)ε\operatorname{Vol}(\overline{K}/\underline{K})/\operatorname{Vol}(K)\leq\varepsilon^{\prime}.

Next, we take n:=3(ε/ε)2n:=\lceil 3(\varepsilon^{\prime}/\varepsilon)^{2}\rceil random samples from K¯K¯\overline{K}\setminus\underline{K}, and we check for each of them whether they are in KK. We denote the fraction of them that is in KK by ξ\xi. This step requires n=O~((ε/ε)2)=O~(ε2(d1)/(d+3))n=\widetilde{O}((\varepsilon^{\prime}/\varepsilon)^{2})=\widetilde{O}(\varepsilon^{-2(d-1)/(d+3)}) membership oracle queries too.

Finally, we output V~=Vol(K)+ξVol(K¯K¯)\widetilde{V}=\operatorname{Vol}(K)+\xi\operatorname{Vol}(\overline{K}\setminus\underline{K}). We observe that

𝔼[V~]=Vol(K¯)+𝔼[ξ]Vol(K¯K¯)=Vol(K¯)+Vol(KK¯)Vol(K¯K¯)Vol(K¯K¯)=Vol(K),\mathbb{E}[\widetilde{V}]=\operatorname{Vol}(\underline{K})+\mathbb{E}[\xi]\operatorname{Vol}(\overline{K}\setminus\underline{K})=\operatorname{Vol}(\underline{K})+\frac{\operatorname{Vol}(K\setminus\underline{K})}{\operatorname{Vol}(\overline{K}\setminus\underline{K})}\operatorname{Vol}(\overline{K}\setminus\underline{K})=\operatorname{Vol}(K),

and

Var[V~]=Vol(K¯K¯)2Var[ξ](εVol(K))2n(εVol(K))23.\operatorname{Var}[\widetilde{V}]=\operatorname{Vol}(\overline{K}\setminus\underline{K})^{2}\operatorname{Var}[\xi]\leq\frac{(\varepsilon^{\prime}\cdot\operatorname{Vol}(K))^{2}}{n}\leq\frac{(\varepsilon\cdot\operatorname{Vol}(K))^{2}}{3}.

We conclude with Chebyshev’s inequality that

[|V~Vol(K)|>εVol(K)]Var[V~](εVol(K))213.\mathbb{P}\left[\left|\widetilde{V}-\operatorname{Vol}(K)\right|>\varepsilon\operatorname{Vol}(K)\right]\leq\frac{\operatorname{Var}[\widetilde{V}]}{(\varepsilon\cdot\operatorname{Vol}(K))^{2}}\leq\frac{1}{3}.\qed

In the quantum case, we speed up the sampling phase from the randomized algorithm by replacing it with a quantum primitive known as amplitude estimation [BHMT02]. We attain a quadratic speed-up of this step, and the claimed complexity follows after rebalancing the costs of the different steps.

Theorem 3.5 (Quantum volume estimation).

Let d2d\in\mathbb{N}_{\geq 2}, R1R\geq 1, ε(0,1)\varepsilon\in(0,1), and KdK\subseteq\mathbb{R}^{d} convex such that BdKRBdB_{d}\subseteq K\subseteq RB_{d}. We can compute V~0\widetilde{V}\geq 0 such that |V~Vol(K)|εVol(K)|\widetilde{V}-\operatorname{Vol}(K)|\leq\varepsilon\operatorname{Vol}(K) with probability at least 2/32/3, using a quantum algorithm that makes O~(ε(d1)/(d+1))\widetilde{O}(\varepsilon^{-(d-1)/(d+1)}) queries to OKO_{K}.

Proof.

Similarly to the proof of Theorem 3.4, we start by constructing an ε\varepsilon^{\prime}-kernel K¯K\underline{K}\subseteq K using Corollary 3.3, and then use Proposition 2.8 to obtain K¯K\overline{K}\supseteq K such that Vol(K¯K¯)εVol(K)\operatorname{Vol}(\overline{K}\setminus\underline{K})\leq\varepsilon^{\prime}\operatorname{Vol}(K). In contrast to Theorem 3.4, though, we choose

ε=(1+ε2d+1)1d14d(d+1)2Θ(ε2d+1),\varepsilon^{\prime}=\frac{(1+\varepsilon^{\frac{2}{d+1}})^{\frac{1}{d}}-1}{4d(d+1)^{2}}\in\Theta\left(\varepsilon^{\frac{2}{d+1}}\right),

which brings the total number of membership oracle queries in this part to O~((ε)(d1)/2)=O~(ε(d1)/(d+1))\widetilde{O}((\varepsilon^{\prime})^{-(d-1)/2})=\widetilde{O}(\varepsilon^{-(d-1)/(d+1)}).

Next, we use amplitude estimation to find an estimate ξ\xi of Vol(KK¯)/Vol(K¯K¯)\operatorname{Vol}(K\setminus\underline{K})/\operatorname{Vol}(\overline{K}\setminus\underline{K}) up to precision ε/ε\varepsilon/\varepsilon^{\prime}. To that end, we compute the overlap between the uniform superposition on K¯K¯\overline{K}\setminus\underline{K}, and the subspace spanned all the quantum states representing vectors in KK. From [BHMT02], we find that this can be done with success probability at least 8/π2>2/38/\pi^{2}>2/3, and with a total number of queries that satisfies O~(ε/ε)=O~(ε(d1)/(d+1))\widetilde{O}(\varepsilon^{\prime}/\varepsilon)=\widetilde{O}(\varepsilon^{-(d-1)/(d+1)}). Finally, we output V~=Vol(K)+ξVol(K¯K¯)\widetilde{V}=\operatorname{Vol}(K)+\xi\operatorname{Vol}(\overline{K}\setminus\underline{K}), and observe that

|V~Vol(K)|=Vol(K¯K¯)|ξVol(KK¯)Vol(K¯K¯)|εVol(K)εε=εVol(K).|\widetilde{V}-\operatorname{Vol}(K)|=\operatorname{Vol}(\overline{K}\setminus\underline{K})\cdot\left|\xi-\frac{\operatorname{Vol}(K\setminus\underline{K})}{\operatorname{Vol}(\overline{K}\setminus\underline{K})}\right|\leq\varepsilon^{\prime}\cdot\operatorname{Vol}(K)\cdot\frac{\varepsilon}{\varepsilon^{\prime}}=\varepsilon\cdot\operatorname{Vol}(K).\qed

This concludes our discussion of randomized and quantum algorithms estimating the volume of a convex body. In Section 4.2, we prove that these algorithms are essentially optimal.

4 Lower bounds

In this section, we prove matching lower bounds for the computational problems from Definition 2.9. We can always rescale our problems, and so we can freely change the assumption that BdKRBdB_{d}\subseteq K\subseteq RB_{d} to (1/R)BdKBd(1/R)B_{d}\subseteq K\subseteq B_{d}. For ease of notation, we will phrase everything with the latter assumption in mind.

The lower bounds crucially rely on embedding a bitstring on the boundary of the unit ball in dd dimensions. To that end, we let {vj}j=1n\{v_{j}\}_{j=1}^{n} be a net of Sd1S_{d-1}, and for every j[n]j\in[n], we define a spherical cap PjP_{j} around vjv_{j}, i.e., a small region cut off from the unit ball by a hyperplane that is orthogonal to vjv_{j}. We take all the spherical caps to be disjoint and with equal volume, and we let K0K_{0} be the remaining part of the unit ball, as shown in Figure 4.1.

K0K_{0}v1v_{1}P1P_{1}v2v_{2}P2P_{2}v3v_{3}P3P_{3}v4v_{4}P4P_{4}v5v_{5}P5P_{5}v6v_{6}P6P_{6}
Figure 4.1: The lower bound construction in two dimensions with n=6n=6. For any x{0,1}6x\in\{0,1\}^{6}, the convex body KxK_{x} is formed by taking the union of K0K_{0} and all spherical caps PjP_{j} if and only if the corresponding bit xjx_{j} is 11.

The core idea is to associate a convex body KxK_{x} to every bitstring x{0,1}nx\in\{0,1\}^{n}, defined as the union of K0K_{0}, and all the spherical caps PjP_{j} if and only if xj=1x_{j}=1. We then immediately observe that Vol(Kx)=Vol(K0)+|x|Vol(P1)\operatorname{Vol}(K_{x})=\operatorname{Vol}(K_{0})+|x|\operatorname{Vol}(P_{1}), and Vol(KxΔKy)=|xy|Vol(P1)\operatorname{Vol}(K_{x}\Delta K_{y})=|x\oplus y|\operatorname{Vol}(P_{1}).

Intuitively, now, if we find an approximation to the convex body KxK_{x}, then we also find an approximation to the bitstring xx. Similarly, if we find a sufficiently precise estimate of KxK_{x}’s volume, we also obtain a good approximation of xx’s Hamming weight. In short, we can reduce the bitstring recovery problem to convex set estimation, and the Hamming weight estimation problem to volume estimation, and hence lower bounds on the former imply lower bounds on the latter.

4.1 Query complexity of bitstring problems

We first recall the bitstring recovery and the Hamming weight estimation problems. These are well-studied in the existing literature, albeit a bit scattered, and we gather the existing results in two concise theorem statements. The proofs can be found in Appendix A.

Theorem 4.1 (Bitstring recovery problem).

Let nn\in\mathbb{N}, and let x{0,1}nx\in\{0,1\}^{n} be a bitstring that we can access through bit queries. Suppose we wish to output a bitstring y{0,1}ny\in\{0,1\}^{n} such that |xy|n/4|x\oplus y|\leq n/4. The query complexities for this problem are Θ(n)\Theta(n) in the deterministic, randomized and quantum setting.

Next, we consider the Hamming weight estimation problem. This problem is sometimes also referred to as the approximate counting problem. In short, given a bitstring of length nn, we wish to estimate the number of 11’s it contains up to some additive precision kk. Its hardness is also well-understood – we gather the previous known results in Theorem 4.2.

Theorem 4.2 (Hamming weight estimation problem).

Let nn\in\mathbb{N} and let x{0,1}nx\in\{0,1\}^{n} be a bitstring that we can access through bit queries. Let kk\in\mathbb{N} be such that 1kn/41\leq k\leq n/4. Suppose we wish to output w[n]0w\in[n]_{0} such that ||x|w|k||x|-w|\leq k. The query complexities for this problem are Θ(n)\Theta(n) in the deterministic setting, Θ(min{(n/k)2,n})\Theta(\min\{(n/k)^{2},n\}) in the randomized setting, and Θ(n/k)\Theta(n/k) in the quantum setting.

4.2 Reduction to convex set estimation and volume estimation

We start by formalizing the concept of a spherical cap, and provide a visualization in Figure 4.2.

Definition 4.3 (Spherical cap).

Let dd\in\mathbb{N}, vSd1v\in S_{d-1} and r>0r>0. Let Sv={uSd1:uv<r}S_{v}=\{u\in S_{d-1}:\left\|u-v\right\|<r\}, and let PvP_{v} be its convex hull. Then PvP_{v} is the spherical cap around vv with radius rr.

SvS_{v}PvP_{v}rrvv
Figure 4.2: The shaded region is a spherical cap around vv with radius rr.

Next, we prove some properties of spherical caps.

Lemma 4.4 (Properties of spherical caps).

Let dd\in\mathbb{N}, v,wSd1v,w\in S_{d-1}, r>0r>0, and PvP_{v} and PwP_{w} the spherical caps with radius rr around vv and ww, respectively. Then,

  1. 1.

    Pv={xBd:xTv>1r2/2}P_{v}=\{x\in B_{d}:x^{T}v>1-r^{2}/2\}.

  2. 2.

    If vw2r\left\|v-w\right\|\geq 2r, then PvP_{v} and PwP_{w} are disjoint.

  3. 3.

    Vol(Pv)Θ(rd+1)\operatorname{Vol}(P_{v})\in\Theta(r^{d+1}), in the limit where dd is fixed and r0r\downarrow 0.

Proof.

For the first claim, observe that xSvx\in S_{v} if and only if xTv=(x2+v2xv2)/2>1r2/2x^{T}v=(\left\|x\right\|^{2}+\left\|v\right\|^{2}-\left\|x-v\right\|^{2})/2>1-r^{2}/2. Consequently, xPvx\in P_{v} if and only if it is inside the unit ball and satisfies this inequality.

For the second claim, we prove the contrapositive, i.e., we prove that zPvPwvw<2rz\in P_{v}\cap P_{w}\Rightarrow\left\|v-w\right\|<2r. To that end, let zPvPwz\in P_{v}\cap P_{w}. Then, using that z1\left\|z\right\|\leq 1 and Cauchy-Schwarz’s inequality, we obtain

vw2=2v2+2w2v+w24z2v+w24(zT(v+w))2<4(2r2)24r2.\left\|v-w\right\|^{2}=2\left\|v\right\|^{2}+2\left\|w\right\|^{2}-\left\|v+w\right\|^{2}\leq 4-\left\|z\right\|^{2}\left\|v+w\right\|^{2}\leq 4-(z^{T}(v+w))^{2}<4-(2-r^{2})^{2}\leq 4r^{2}.

Finally, computing the volume yields

Vol(Pv)=Γd11r2/21(1x2)d12dx=Γd10r2/2(x(2x))d12dxΘ(rd+1).\operatorname{Vol}(P_{v})=\Gamma_{d-1}\int_{1-r^{2}/2}^{1}(1-x^{2})^{\frac{d-1}{2}}\;\mathrm{d}x=\Gamma_{d-1}\int_{0}^{r^{2}/2}(x(2-x))^{\frac{d-1}{2}}\;\mathrm{d}x\in\Theta(r^{d+1}).\qed

We now use the above properties to compute how many spherical caps we can pack on the surface of a unit ball.

Lemma 4.5 (Packing of spherical caps).

Let d2d\in\mathbb{N}_{\geq 2} and let δ>0\delta>0. Let {vj}j=1n\{v_{j}\}_{j=1}^{n} be a δ\delta-net, for each j[n]j\in[n] let Pj:=PvjP_{j}:=P_{v_{j}} be a spherical cap with radius δ/2\delta/2, and let P={Pj}j=1nP=\{P_{j}\}_{j=1}^{n}. Then, n=Θ(δ(d1))n=\Theta(\delta^{-(d-1)}), all the spherical caps are disjoint, and the total volume of the spherical caps is Vol(P):=nVol(P1)=Θ(δ2)\operatorname{Vol}(P):=n\operatorname{Vol}(P_{1})=\Theta(\delta^{2}).

Proof.

The claims follow immediately from Proposition 2.4 and Lemma 4.4. ∎

Finally, we show how embedding a bitstring in these spherical caps naturally leads to a lower bound on the query complexity of the volume estimation and the convex set estimation problems.

Theorem 4.6.

Let d2d\in\mathbb{N}_{\geq 2} and ε>0\varepsilon>0. Then, any algorithm 𝒜\mathcal{A} that solves the convex set estimation problem or the volume estimation problem, must make a number of queries that satisfies the complexities listed in Table 1.1.

Proof.

We start with the lower bounds on the convex set estimation problems. To that end, suppose that an algorithm 𝒜\mathcal{A} solves it with relative Nikodym distance ε\varepsilon.

For every δ>0\delta>0, we choose a particular δ\delta-net {vj}j=1n\{v_{j}\}_{j=1}^{n}, and then we let {Pj}j=1n\{P_{j}\}_{j=1}^{n} and PP be as in Lemma 4.5. Observe that Vol(P)Θ(δ2)\operatorname{Vol}(P)\in\Theta(\delta^{2}). We now choose the smallest δ>0\delta>0 such that Vol(P)8εΓd\operatorname{Vol}(P)\geq 8\varepsilon\Gamma_{d}, which for small enough ε>0\varepsilon>0 is guaranteed to be well-defined. Then δΘ(ε)\delta\in\Theta(\sqrt{\varepsilon}), which implies that nΘ(ε(d1)/2)n\in\Theta(\varepsilon^{-(d-1)/2}).

Now, for any bitstring x{0,1}nx\in\{0,1\}^{n}, we define a convex body KxK_{x} by

Kx=K0j=1xj=1nPj,K_{x}=K_{0}\cup\bigcup_{\begin{subarray}{c}j=1\\ x_{j}=1\end{subarray}}^{n}P_{j}, (2)

Then, given any bitstring x{0,1}nx\in\{0,1\}^{n}, 𝒜\mathcal{A} outputs K~\widetilde{K}, an approximation of KxK_{x} with relative precision ε\varepsilon. Next, we can define a bitstring z{0,1}nz\in\{0,1\}^{n} such that

Vol(K~ΔKz)=min{Vol(K~ΔKy):y{0,1}n}Vol(K~ΔKx)εVol(Kx)εΓd.\operatorname{Vol}(\widetilde{K}\Delta K_{z})=\min\{\operatorname{Vol}(\widetilde{K}\Delta K_{y}):y\in\{0,1\}^{n}\}\leq\operatorname{Vol}(\widetilde{K}\Delta K_{x})\leq\varepsilon\operatorname{Vol}(K_{x})\leq\varepsilon\Gamma_{d}.

Then, we obtain that

|xz|=Vol(KzΔKx)Vol(P1)n(Vol(K~ΔKz)+Vol(K~ΔKx))Vol(P)n(εΓd+εΓd)8εΓd=n4.|x\oplus z|=\frac{\operatorname{Vol}(K_{z}\Delta K_{x})}{\operatorname{Vol}(P_{1})}\leq\frac{n(\operatorname{Vol}(\widetilde{K}\Delta K_{z})+\operatorname{Vol}(\widetilde{K}\Delta K_{x}))}{\operatorname{Vol}(P)}\leq\frac{n(\varepsilon\Gamma_{d}+\varepsilon\Gamma_{d})}{8\varepsilon\Gamma_{d}}=\frac{n}{4}.

Finally, we observe that any call to the membership oracle of KxK_{x} can be simulated with at most one query to the bitstring xx. Thus, according to Theorem 4.1, 𝒜\mathcal{A} must make at least Ω(n)=Ω(ε(d1)/2)\Omega(n)=\Omega(\varepsilon^{-(d-1)/2}) queries to the membership oracle of KxK_{x}, in the deterministic, randomized and quantum settings. Finally, since the ε\varepsilon-kernel construction problem is harder than the ε\varepsilon-Nikodym construction problem, the lower bounds from the first two columns in Table 1.1 follow.

It remains to prove the lower bounds for volume estimation. To that end, suppose that 𝒜\mathcal{A} solves the volume estimation problem up to precision ε\varepsilon. Again for every δ>0\delta>0 we choose a particular δ\delta-net {vj}j=1n\{v_{j}\}_{j=1}^{n}, and we let {Pj}j=1n\{P_{j}\}_{j=1}^{n} and PP as in Lemma 4.5, which means that Vol(P)Θ(δ2)\operatorname{Vol}(P)\in\Theta(\delta^{2}). Next, we let ε>0\varepsilon^{\prime}>0 arbitrarily and choose the smallest δ>0\delta>0 such that Vol(P)8εΓd\operatorname{Vol}(P)\geq 8\varepsilon^{\prime}\Gamma_{d}. For small enough ε\varepsilon^{\prime}, this is well-defined, and we find that δΘ(ε)\delta\in\Theta(\sqrt{\varepsilon^{\prime}}), from which we find that nΘ((ε)(d1)/2)n\in\Theta((\varepsilon^{\prime})^{-(d-1)/2}).

We let x{0,1}nx\in\{0,1\}^{n} and KxK_{x} as in Eq. 2. Then,

Vol(Kx)=Vol(K0)+|x|Vol(P1).\operatorname{Vol}(K_{x})=\operatorname{Vol}(K_{0})+|x|\operatorname{Vol}(P_{1}).

𝒜\mathcal{A} outputs an estimate V~\widetilde{V} of Vol(Kx)\operatorname{Vol}(K_{x}), such that |V~Vol(Kx)|εVol(Kx)εΓd|\widetilde{V}-\operatorname{Vol}(K_{x})|\leq\varepsilon\operatorname{Vol}(K_{x})\leq\varepsilon\Gamma_{d}. Hence, we can compute w=(V~Vol(K0))/Vol(P1)w=(\widetilde{V}-\operatorname{Vol}(K_{0}))/\operatorname{Vol}(P_{1}), and we observe that

|w|x||=|V~Vol(K0)Vol(P1)Vol(Kx)Vol(K0)Vol(P1)|=n|V~Vol(Kx)|Vol(P)nεΓd8εΓd=nε8ε=:k.|w-|x||=\left|\frac{\widetilde{V}-\operatorname{Vol}(K_{0})}{\operatorname{Vol}(P_{1})}-\frac{\operatorname{Vol}(K_{x})-\operatorname{Vol}(K_{0})}{\operatorname{Vol}(P_{1})}\right|=\frac{n|\widetilde{V}-\operatorname{Vol}(K_{x})|}{\operatorname{Vol}(P)}\leq\frac{n\varepsilon\Gamma_{d}}{8\varepsilon^{\prime}\Gamma_{d}}=\frac{n\varepsilon}{8\varepsilon^{\prime}}=:k.

Finally, similarly as before, since we can simulate any query to a membership oracle to KxK_{x} by one query to the bitstring xx, we can use the lower bounds from Theorem 4.2 to lower bound the query complexity of 𝒜\mathcal{A}.

For the lower bounds of Theorem 4.2 to apply, we have to choose ε>0\varepsilon^{\prime}>0 as a function of ε\varepsilon such that 1kn/41\leq k\leq n/4. To that end, we let =ε/2\ell=\varepsilon/2 and u=nε/8u=n\varepsilon/8, and observe that we must choose ε[,u]\varepsilon^{\prime}\in[\ell,u]. We choose ε\varepsilon^{\prime} as \ell, u\sqrt{\ell u} and uu, in the deterministic, randomized, and quantum settings, respectively. We remark in the randomized setting that n/k=8ε/εΘ(n)n/k=8\varepsilon^{\prime}/\varepsilon\in\Theta(\sqrt{n}), and so the lower bound from Theorem 4.2 becomes Ω(n)\Omega(n). Similarly, in the quantum case, we have kΘ(1)k\in\Theta(1), and so the lower bound also becomes Ω(n)\Omega(n). It remains to plug in the asymptotic scaling of nn, from which we obtain the following lower bounds for the volume estimation problem:

Deterministic: Ω(n)=Ω((ε)d12)=Ω(εd12)\displaystyle\Omega(n)=\Omega((\varepsilon^{\prime})^{-\frac{d-1}{2}})=\Omega(\varepsilon^{-\frac{d-1}{2}})
Randomized: Ω(n)=Ω(n4d+3nd1d+3)=Ω((ε)2(d1)d+3(ε/ε)2(d1)d+3)=Ω(ε2(d1)d+3).\displaystyle\Omega(n)=\Omega(n^{\frac{4}{d+3}}\cdot n^{\frac{d-1}{d+3}})=\Omega((\varepsilon^{\prime})^{-\frac{2(d-1)}{d+3}}\cdot(\varepsilon^{\prime}/\varepsilon)^{\frac{2(d-1)}{d+3}})=\Omega(\varepsilon^{-\frac{2(d-1)}{d+3}}).
Quantum: Ω(n)=Ω(n2d+1nd1d+1)=Ω((ε)d1d+1(ε/ε)d1d+1)=Ω(εd1d+1).\displaystyle\Omega(n)=\Omega(n^{\frac{2}{d+1}}\cdot n^{\frac{d-1}{d+1}})=\Omega((\varepsilon^{\prime})^{-\frac{d-1}{d+1}}\cdot(\varepsilon^{\prime}/\varepsilon)^{\frac{d-1}{d+1}})=\Omega(\varepsilon^{-\frac{d-1}{d+1}}).\qed

Acknowledgements

We would like to thank anonymous reviewers for directing our attention to the existing literature on ε\varepsilon-kernels, and many helpful comments on the presentation of the result. SA was supported in part by the European QuantERA project QOPT (ERA-NET Cofund 2022-25), the French PEPR integrated projects EPiQ (ANR-22-PETQ-0007) and HQI (ANR-22-PNCQ-0002), and the French ANR project QUOPS (ANR-22-CE47-0003-01). AC was supported by a Simons-CIQC postdoctoral fellowship through NSF QLCI Grant No. 2016245.

References

  • [AHP24] Pankaj K Agarwal and Sariel Har-Peled. Computing instance-optimal kernels in two dimensions. Discrete & Computational Geometry, pages 1–28, 2024.
  • [AHPV04] Pankaj K Agarwal, Sariel Har-Peled, and Kasturi R Varadarajan. Approximating extent measures of points. Journal of the ACM (JACM), 51(4):606–635, 2004.
  • [Amb00] Andris Ambainis. Quantum lower bounds by quantum arguments. In Proceedings of the 32nd annual ACM symposium on Theory of computing, pages 636–643, 2000.
  • [BB20] Shalev Ben-David and Eric Blais. A tight composition theorem for the randomized query complexity of partial functions. In 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), pages 240–246. IEEE, 2020.
  • [BBC+01] Robert Beals, Harry Buhrman, Richard Cleve, Michele Mosca, and Ronald de Wolf. Quantum lower bounds by polynomials. Journal of the ACM (JACM), 48(4):778–797, 2001.
  • [BF87] Imre Bárány and Zoltán Füredi. Computing the volume is difficult. Discret. Comput. Geom., 2:319–326, 1987.
  • [BHMT02] Gilles Brassard, Peter Hoyer, Michele Mosca, and Alain Tapp. Quantum amplitude amplification and estimation. Contemporary Mathematics, 305:53–74, 2002.
  • [BI75] E.M. Bronshteyn and L.D. Ivanov. The approximation of convex sets by polyhedra. Siberian Mathematical Journal, 16(5):852–853, 1975.
  • [BR16] Nikolay Baldin and Markus Reiß. Unbiased estimation of the volume of a convex body. Stochastic Processes and their Applications, 126(12):3716–3732, 2016.
  • [Bru16] Victor-Emmanuel Brunel. Adaptive estimation of convex and polytopal density support. Probability Theory and Related Fields, 164(1):1–16, 2016.
  • [Bru18] Victor-Emmanuel Brunel. Methods for estimation of convex sets. Statistical Science, 33(4):615–632, 2018.
  • [CCH+23] Shouvanik Chakrabarti, Andrew M. Childs, Shih-Han Hung, Tongyang Li, Chunhao Wang, and Xiaodi Wu. Quantum algorithm for estimating volumes of convex bodies. ACM Transactions on Quantum Computing, 4(3):1–60, 2023.
  • [CH23] Arjan Cornelissen and Yassine Hamoudi. A sublinear-time quantum algorithm for approximating partition functions. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 1245–1264. SIAM, 2023.
  • [Cha06] Timothy M Chan. Faster core-set constructions and data-stream algorithms in fixed dimensions. Computational Geometry, 35(1-2):20–35, 2006.
  • [Che21] Yuansi Chen. An almost constant lower bound of the isoperimetric coefficient in the KLS conjecture. Geometric and Functional Analysis, 31:34–61, 2021.
  • [CHJ22] Arjan Cornelissen, Yassine Hamoudi, and Sofiene Jerbi. Near-optimal quantum algorithms for multivariate mean estimation. In Proceedings of the 54th Annual ACM SIGACT Symposium on Theory of Computing, pages 33–43, 2022.
  • [CV18] Ben Cousins and Santosh Vempala. Gaussian cooling and O(n3)O^{*}(n^{3}) algorithms for volume and Gaussian volume. SIAM Journal on Computing, 47(3):1237–1273, 2018.
  • [DFK91] Martin Dyer, Alan Frieze, and Ravi Kannan. A random polynomial-time algorithm for approximating the volume of convex bodies. Journal of the ACM (JACM), 38(1):1–17, 1991.
  • [Dud74] R.M. Dudley. Metric entropy of some classes of sets with differentiable boundaries. Journal of Approximation Theory, 10(3):227–236, 1974.
  • [Ele86] György Elekes. A geometric inequality and the complexity of computing volume. Discrete & Computational Geometry, 1:289–292, 1986.
  • [FGGS99] Edward Farhi, Jeffrey Goldstone, Sam Gutmann, and Michael Sipser. Bound on the number of functions that can be distinguished with kk quantum queries. Physical Review A, 60(6):4331, 1999.
  • [GLS88] Martin Grötschel, László Lovász, and Alexander Schrijver. Geometric algorithms and combinatorial optimization. Springer-Verlag Berlin Heidelberg, 1 edition, 1988.
  • [HJ19] Sariel Har-Peled and Mitchell Jones. Proof of Dudley’s convex approximation. arXiv preprint arXiv:1912.01977, 2019.
  • [JLLV21] He Jia, Aditi Laddha, Yin Tat Lee, and Santosh Vempala. Reducing isotropy and volume to KLS: an O(n3ψ2)O^{*}(n^{3}\psi^{2}) volume algorithm. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, pages 961–974, 2021.
  • [JLV22] Arun Jambulapati, Yin Tat Lee, and Santosh S. Vempala. A slightly improved bound for the KLS constant. arXiv preprint arXiv:2208.11644, 2022.
  • [KL22] Bo’az Klartag and Joseph Lehec. Bourgain’s slicing problem and KLS isoperimetry up to polylog. Geometric and Functional Analysis, 32(5):1134–1159, 2022.
  • [Kla23] Bo’az Klartag. Logarithmic bounds for isoperimetry and slices of convex sets. arXiv preprint arXiv:2303.14938, 2023.
  • [LV06] László Lovász and Santosh Vempala. Simulated annealing in convex bodies and an O(n4)O^{*}(n^{4}) volume algorithm. Journal of Computer and System Sciences, 72(2):392–417, 2006.
  • [NC10] Michael A. Nielsen and Isaac L. Chuang. Quantum computation and quantum information. Cambridge University Press, 2010.
  • [NW99] Ashwin Nayak and Felix Wu. The quantum query complexity of approximating the median and related statistics. In Proceedings of the thirty-first annual ACM symposium on Theory of computing, pages 384–393, 1999.
  • [RV08] Luis Rademacher and Santosh Vempala. Dispersion of mass and the complexity of randomized geometric algorithms. Advances in Mathematics, 219(3):1037–1069, 2008.
  • [Sch94] Carsten Schütt. Random polytopes and affine surface area. Mathematische Nachrichten, 170(1):227–249, 1994.
  • [Sim03] Miklós Simonovits. How to compute the volume in high dimension? Mathematical programming, 97:337–374, 2003.
  • [YAPV08] Hai Yu, Pankaj K Agarwal, Raghunath Poreddy, and Kasturi R Varadarajan. Practical methods for shape fitting and kinetic data structures using coresets. Algorithmica, 52:378–402, 2008.

Appendix A Query complexities of bitstring problems

In this appendix, we provide the proofs to the query complexities of the bitstring problems that we use in the lower bound constructions in Section 4. We restate the theorems for convenience.

See 4.1

Proof.

The (deterministic) algorithm is trivial – simply query all the entries of xx and output the bitstring exactly. Since the randomized and quantum settings can simulate the deterministic setting, this provides the upper bounds on the query complexities.

It remains to prove that this algorithm is optimal in all three models up to constant multiplicative factors. Since we can simulate any deterministic and randomized algorithm in the quantum setting, proving hardness in the latter suffices. An information-theoretic argument tells us that quantumly we need at least Ω(n)\Omega(n) queries to output a bitstring that differs from xx in at most n/4n/4 positions. The core technique stems from [FGGS99], and all remaining details can for instance be found in [CHJ22, Lemma 4.6]. ∎

See 4.2

Proof.

We first prove the upper bounds. The deterministic upper bound is trivial. In the randomized setting, we need to show that the query complexity is O((n/k)2)O((n/k)^{2}). To that end, suppose that we sample 3(n/k)23(n/k)^{2} bits at random, with replacement, and output nn times the fraction of 11’s observed. Let XjX_{j} be the random variable describing the output of the jjth sample, and we write the output of the algorithm as w:=n/3(k/n)2j=13(n/k)2Xjw:=n/3\cdot(k/n)^{2}\cdot\sum_{j=1}^{3(n/k)^{2}}X_{j}. Then, we trivially find that 𝔼[w]=|x|\mathbb{E}[w]=|x|, and its variance can be bounded as

Var[w]=Var[n3(kn)2j=13(n/k)2Xj]=n29(kn)43(nk)2Var[X1]=13k2Var[X1]13k2,\operatorname{Var}[w]=\operatorname{Var}\left[\frac{n}{3}\left(\frac{k}{n}\right)^{2}\cdot\sum_{j=1}^{3(n/k)^{2}}X_{j}\right]=\frac{n^{2}}{9}\left(\frac{k}{n}\right)^{4}\cdot 3\left(\frac{n}{k}\right)^{2}\operatorname{Var}[X_{1}]=\frac{1}{3}k^{2}\operatorname{Var}[X_{1}]\leq\frac{1}{3}k^{2},

and so by Chebyshev’s inequality, we have

[|w|x||k]Var[w]k213.\mathbb{P}[|w-|x||\geq k]\leq\frac{\operatorname{Var}[w]}{k^{2}}\leq\frac{1}{3}.

Finally, quantumly, the algorithm is known as the quantum approximate counting algorithm [BHMT02], and the upper bound follows directly.

Then, it remains to prove the lower bounds. Deterministically, the problem is easiest when kk is largest, so it suffices to prove the lower bound when k=n/4k=n/4. Once we have queried n/21n/2-1 bits of the bitstring – let’s say we observed ww ones – there are still n/2+1n/2+1 bits left to query. Then, we know that the true Hamming weight must be between ww and w+n/2+1w+n/2+1. Since this interval is larger than 2k2k, we cannot guarantee that whatever number we output is at most kk away from the true Hamming weight. Thus, we need to make at least n/2Ω(n)n/2\in\Omega(n) queries.

In the randomized setting, we use [BB20, Lemma 26]. If k=nk=\sqrt{n}, we need at least Ω(n)\Omega(n) queries. Since the problem becomes more difficult when kk decreases, this proves the lower bound for all knk\leq\sqrt{n}. On the other hand, if k>nk>\sqrt{n}, then we define nΘ((n/k)2)n^{\prime}\in\Theta((n/k)^{2}). For any bitstring x{0,1}nx^{\prime}\in\{0,1\}^{n^{\prime}}, we define a bitstring xx of length Θ(n)\Theta(n) by duplicating all the bits of xx^{\prime} a total of n/nΘ(k2/n)n/n^{\prime}\in\Theta(k^{2}/n) times. Then, a Hamming weight estimation algorithm with precision kk will find an approximation of the Hamming weight of xx with precision kk, which means that it outputs an approximation of the Hamming weight of xx^{\prime} with precision Θ(k/(k2/n))=Θ(n/k)=Θ(n)\Theta(k/(k^{2}/n))=\Theta(n/k)=\Theta(\sqrt{n^{\prime}}). Thus, this algorithm must query xx at least Ω(n)=Ω((n/k)2)\Omega(n^{\prime})=\Omega((n/k)^{2}) times.

Finally, in the quantum setting, the result follows easily from [Amb00]. For k=1k=1, the hardness already follows from the majority function, whose hardness was proved in [BBC+01], and the hardness of the general case was basically proved in [NW99], even though not phrased as such. ∎