This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Near Optimal Alphabet-Soundness Tradeoff PCPs

Dor Minzer Department of Mathematics, Massachusetts Institute of Technology, Cambridge, USA. Supported by NSF CCF award 2227876 and NSF CAREER award 2239160.    Kai Zhe Zheng Department of Mathematics, Massachusetts Institute of Technology, Cambridge, USA. Supported by the NSF GRFP DGE-2141064.
Abstract

We show that for all ε>0\varepsilon>0, for sufficiently large prime power qq\in\mathbb{N}, for all δ>0\delta>0, it is NP-hard to distinguish whether a 22-Prover-11-Round projection game with alphabet size qq has value at least 1δ1-\delta, or value at most 1/q1ε1/q^{1-\varepsilon}. This establishes a nearly optimal alphabet-to-soundness tradeoff for 22-query PCPs with alphabet size qq, improving upon a result of Chan [Cha16]. Our result has the following implications:

  1. 1.

    Near optimal hardness for Quadratic Programming: it is NP-hard to approximate the value of a given Boolean Quadratic Program within factor (logn)1o(1)(\log n)^{1-o(1)} under quasi-polynomial time reductions. This result improves a result of Khot and Safra [KS13] and nearly matches the performance of the best known approximation algorithm [Meg01, NRT99, CW04] that achieves a factor of O(logn)O(\log n).

  2. 2.

    Bounded degree 22-CSP’s: under randomized reductions, for sufficiently large d>0d>0, it is NP-hard to approximate the value of 22-CSPs in which each variable appears in at most dd constraints within factor (1o(1))d2(1-o(1))\frac{d}{2}, improving upon a recent result of Lee and Manurangsi [LM23].

  3. 3.

    Improved hardness results for connectivity problems: using results of Laekhanukit [Lae14] and Manurangsi [Man19], we deduce improved hardness results for the Rooted kk-Connectivity Problem, the Vertex-Connectivity Survivable Network Design Problem and the Vertex-Connectivity kk-Route Cut Problem.

1 Introduction

The PCP theorem is a fundamental result in theoretical computer science with many equivalent formulations [FGL+91, AS03, ALM+92]. One of the formulations asserts that there exists ε>0\varepsilon>0 such that given a satisfiable 33-SAT formula ϕ\phi, it is NP-hard to find an assignment that satisfies at least (1ε)(1-\varepsilon) fraction of the constraints. The PCP theorem has a myriad of applications within theoretical computer science, and of particular interest to this paper are applications of PCP to hardness of approximation.

The vast majority of hardness of approximation result are proved via reductions from the PCP theorem above. Oftentimes, to get a strong hardness of approximation result, one must first amplify the basic PCP theorem above into a result with stronger parameters [Hås01, Hås96, Fei98, KP06] (see [Tre14] for a survey). To discuss these parameters, it is often convenient to view the PCP through the problem of 22-Prover-11-Round Games, which we define next.111Strictly speaking, the notion below is referred to in the literature as projection 22-Prover-11-Round games. We omit the more general definition as we do not discuss non-projection games in this paper.

Definition 1.1.

An instance Ψ\Psi of 22-Prover-11-Round Games consists of a bipartite graph G=(LR,E)G=(L\cup R,E), alphabets ΣL\Sigma_{L} and ΣR\Sigma_{R} and a collection of constraints Φ={ϕe}eE\Phi=\{\phi_{e}\}_{e\in E}, which for each edge eEe\in E specifies a constraint map ϕe:ΣLΣR\phi_{e}\colon\Sigma_{L}\to\Sigma_{R}.

  1. 1.

    The alphabet size of Ψ\Psi is defined to be |ΣL|+|ΣR||\Sigma_{L}|+|\Sigma_{R}|.

  2. 2.

    The value of Ψ\Psi is defined to be the maximum fraction of edges eEe\in E that can be satisfied by any assignment. That is,

    𝗏𝖺𝗅(Ψ)=maxAL:LΣLAR:RΣR|{e=(u,v)E|ϕe(AL(u))=AR(v)}||E|.{\sf val}(\Psi)=\max_{\begin{subarray}{c}A_{L}\colon L\to\Sigma_{L}\\ A_{R}\colon R\to\Sigma_{R}\end{subarray}}\frac{|\{e=(u,v)\in E~{}|~{}\phi_{e}(A_{L}(u))=A_{R}(v)\}|}{|E|}.

The combinatorial view of 22-Prover-11-Round Games has its origins in an equivalent, active view in terms of a game between a verifier and two all powerful provers, which is sometimes more intuitive. The verifier and the two provers have access to an instance Ψ\Psi of 22-Prover-11-Round Games, and the provers may agree beforehand on a strategy; after this period they are not allowed to communicate. The verifier then picks a random edge, e=(u,v)e=(u,v), from the 22-Prover-11-Round game, sends uu to the first prover, sends vv to the second prover, receives a label in response from each one of them, and finally checks that the labels satisfy the constraint ϕe\phi_{e}. If so, then the verifier accepts. It is easy to see that the value of the 22-Prover-11-Round game is equal to the acceptance probability of the verifier under the best strategy of the provers. This view will be useful for us later.

In the language of 22-Prover-11-Round Games, the majority of hardness of approximation results are proved by combining the basic PCP theorem [FGL+91, AS03, ALM+92] with Raz’s parallel repetition theorem [Raz98], which together imply the following result:

Theorem 1.2.

There exists γ>0\gamma>0 such that for sufficiently large RR, given a 22-Prover-11-Round game Ψ\Psi with alphabet size RR, it is NP-hard to distinguish between the following two cases:

  1. 1.

    YES case: 𝗏𝖺𝗅(Ψ)=1{\sf val}(\Psi)=1.

  2. 2.

    NO case: 𝗏𝖺𝗅(Ψ)1Rγ{\sf val}(\Psi)\leqslant\frac{1}{R^{\gamma}}.

For many applications, one only requires that the soundness error of the PCP is small. Namely, that 𝗏𝖺𝗅(Ψ){\sf val}(\Psi) is arbitrarily small in the “NO case”. For certain applications however, more is required: not only must the soundness error be small – but it must also be small in terms of the alphabet size. The tradeoff between the soundness error of the PCP and the alphabet size of the PCP is the main focus of this paper.

With respect to this tradeoff, it is clear that the best result one may hope for in Theorem 1.2 is γ=1o(1)\gamma=1-o(1) since a random assignment to Ψ\Psi satisfies, in expectation, at least 1R\frac{1}{R} fraction of the constraints. In terms of results, combining the PCP theorem with Raz’s parallel repetition theorem gives γ>0\gamma>0 that is an absolute, but tiny constant. Towards a stronger tradeoff, Khot and Safra [KS13] showed that Theorem 1.2 holds for γ=1/6\gamma=1/6 with imperfect completeness (i.e., 𝗏𝖺𝗅(Ψ)1o(1){\sf val}(\Psi)\geqslant 1-o(1) instead of 𝗏𝖺𝗅(Ψ)=1{\sf val}(\Psi)=1 in the YES case). The result of Khot and Safra was improved by Chan [Cha16], who showed (using a completely different set of techniques) that Theorem 1.2 holds for γ=1/2o(1)\gamma=1/2-o(1), again with imperfect completeness.

1.1 Main Results

In this section we explain the main results of this paper.

1.1.1 Near Optimal Alphabet vs Soundness Tradeoff

The main result of this work improves upon all prior results, and shows that one may take γ=1o(1)\gamma=1-o(1) in Theorem 1.2, again with imperfect completeness. Formally, we show:

Theorem 1.3.

For all ε,δ>0\varepsilon,\delta>0, for sufficiently large RR, given a 22-Prover-11-Round game Ψ\Psi, it is NP-hard to distinguish between the following two cases:

  1. 1.

    YES case: 𝗏𝖺𝗅(Ψ)1δ{\sf val}(\Psi)\geqslant 1-\delta.

  2. 2.

    NO case: 𝗏𝖺𝗅(Ψ)1R1ε{\sf val}(\Psi)\leqslant\frac{1}{R^{1-\varepsilon}}.

Theorem 1.3 shows a near optimal tradeoff between the alphabet of a PCP and the alphabet of a PCP, improving upon the result of Chan [Cha16]. Moreover, Theorem 1.3 has several applications to combinatorial optimization problems, which we discuss below. We remark that most of these applications require additional features from the instances produced in Theorem 1.3 which we omit from its formulation for the sake of clarity. For instance, one application requires a good tradeoff between the size of the instance and the size of the alphabet, which our construction achieves (see the discussion following Theorem 1.4). Other applications require the underlying constraint graph to be bounded-degree bi-regular graph, which our construction also achieves (after mild modifications; see Theorem 7.1).

1.1.2 Application: NP-Hardness of Approximating Quadratic Programs

Theorem 1.3 has an application to the hardness of approximating the value of Boolean Quadratic Programming, as we explain next.

An instance of Quadratic programming consists of a quadratic form Q(x)=i,j=1nai,jxixjQ(x)=\sum\limits_{i,j=1}^{n}a_{i,j}x_{i}x_{j} where ai,i=0a_{i,i}=0 for all ii, and one wishes to maximize Q(x)Q(x) over x{1,1}nx\in\{-1,1\}^{n}. This problem is known to have an O(logn)O(\log n) approximation algorithm [Meg01, NRT99, CW04], and is known to be quasi-NP-hard to approximate within factor (logn)1/6o(1)(\log n)^{1/6-o(1)} [ABH+05, KS13]. That is, unless NP has a quasi-polynomial time algorithm, no polynomial time algorithm can approximate Quadratic Programming to within factor (logn)1/6o(1)(\log n)^{1/6-o(1)}. As a first application of Theorem 1.3, we improve the hardness result of Khot and Safra:

Theorem 1.4.

It is quasi-NP-hard to approximate Quadratic Programming to within a factor of (logn)1o(1)(\log n)^{1-o(1)}.

Theorem 1.4 is proved via a connection between 22-Prover-11-Round Games and Quadratic Programming due to Arora, Berger, Hazan, Kindler, and Safra [ABH+05]. This connections requires a good tradeoff between the alphabet size, the soundness error of the PCP, and the size of the PCP. Fortunately, the construction in Theorem 1.4 has a sufficiently good tradeoff between all of these parameters: letting NN be the size of the instance, the alphabet size can be taken to be (logN)1o(1)(\log N)^{1-o(1)} and the soundness error can be taken to be (logN)1+o(1)(\log N)^{-1+o(1)}. 222We remark that the result of Chan [Cha16] does not achieve a good enough trade-off between the alphabet size and the instance size due to the use of the long-code, and therefore it does not yield a strong inapproximability result for Quadratic Programming.

Relevance to the sliding scale conjecture:

It is worth noting that using our techniques, we do not know how to achieve soundness error that is smaller than inversely poly-logarithmic in the instance size. In particular, our techniques have no bearing on the sliding scale conjecture, which is concerned with getting soundness error that is inversely polynomial in the instance size. This seems to be a bottleneck of any PCP construction that is based on the covering property. In fact, assuming ETH, any quasi-polynomial PCP construction achieving soundness error, say, 1/(logN)21/(\log N)^{2} would necessarily need to have almost polynomial alphabet size (since the reduction to Quadratic Solvability would give an algorithm that runs roughly in time exponential in the alphabet size), which is the opposite of what our techniques give. With this in mind, we would like to mention a closely related, recent conjecture made in [CDGT22], which is a sort of a mixture between dd-to-11 games and the sliding scale conjecture. This conjecture is motivated by improved hardness results for densest kk-subgraph style problems, and focuses on the relation between the instance size and the soundness error (allowing the alphabet to be quite large). It may be possible that the ideas from the current paper can help make progress towards this conjecture.

1.1.3 Application: NP-hardness of Approximating Bounded Degree 22-CSPs

Theorem 1.3 has an application to the hardness of approximating the value of 22-CSPs with bounded degree, as we explain next.

An instance Ψ\Psi of 22-CSP, say Ψ=(X,C,Σ)\Psi=(X,C,\Sigma), consists of a set of variables XX, a set of constraints CC and an alphabet Σ\Sigma. Each constraint in CC has the form P(xi,xj)=1P(x_{i},x_{j})=1 where P:Σ×Σ{0,1}P\colon\Sigma\times\Sigma\to\{0,1\} is a predicate (which may be different in distinct constraints). The degree of the instance Ψ\Psi is defined to be the maximum, over variables xXx\in X, of the number of constraints that xx appears in. The goal is to find an assignment A:XΣA\colon X\to\Sigma that satisfies as many of the constraints as possible.

There is a simple d+12\frac{d+1}{2} approximation algorithm for the 22-CSP problem for instances with degree at most dd. Lee and Manurangsi proved a nearly matching (12o(1))d\left(\frac{1}{2}-o(1)\right)d hardness of approximation result assuming the Unique-Games Conjecture [LM23]. Unconditionally, they show the problem to be NP-hard to approximate within factor (13o(1))d\left(\frac{1}{3}-o(1)\right)d under randomized reductions.

Using the ideas of Lee and Manurangsi, our main result implies a nearly matching NP-hardness result for bounded degree 22-CSPs:

Theorem 1.5.

For all η>0\eta>0, for sufficiently large dd, approximating the value of 22-CSPs with degree at most dd within factor (12η)d\left(\frac{1}{2}-\eta\right)d is NP-hard under randomized reductions.

As in [LM23], Theorem 1.5 has a further application to finding independent sets in claw free graphs. A kk-claw K1,kK_{1,k} is the (k+1)(k+1) vertex graph with a center vertex which is connected to all other kk-vertices and has no other edges; a graph GG is said to be kk-claw free if GG does not contain an induced kk-claw graph. There is a polynomial time approximation algorithm for approximating the size of the largest independent set in a given kk-claw free graph GG within factor k2\frac{k}{2} [Ber00, TW23], and a quasi-polynomial time approximation algorithm within factor (13+o(1))k\left(\frac{1}{3}+o(1)\right)k [CGM13]. As in [LM23], using ideas from [DFRR23] Theorem 1.5 implies that for all ε>0\varepsilon>0, for sufficiently large kk, it is NP-hard (under randomized reductions) to approximate the size of the largest independent set in a given kk-claw free graph within factor (14+η)k\left(\frac{1}{4}+\eta\right)k. This improves upon the result of [LM23] who showed that the same result holds assuming the Unique-Games Conjecture.

1.1.4 Application: NP-hardness of Approximating Connectivity Problems

Using ideas of Laekhanukit [Lae14] and the improvements by Manurangsi [Man19], Theorem 1.3 implies improved hardness of approximation results for several graph connectivitiy problems. More specifically, Theorem 1.3 combined with the results of [Man19] implies improvements to each one of the results outlined in table 11 in [Lae14] by a factor of 22 in the exponent - with the exception of Rooted-kk-Connectivity on directed graphs where a factor of 22 improvement is already implied by [Man19]. We briefly discuss the Rooted kk-Connectivity Problem, but defer the reader to [Lae14] for a detailed discussion of the remaining graph connectivity problems.

In the Rooted kk-Connectivity problem there is a graph G=(V,E)G=(V,E), edge costs c:Ec\colon E\to\mathbb{R}, a root vertex rVr\in V and a set of terminals TV{r}T\subseteq V\setminus\{r\}. The goal is to find a sub-graph GG^{\prime} of smallest cost that for each tTt\in T, has at least kk vertex disjoint paths from rr to tt. The problem admits |T||T| trivial approximation algorithm (by applying minimum cost kk-flow algorithm for each vertex in TT), as well as an O(klogk)O(k\log k) approximation algorithm [Nut12].

Using the ideas of [Lae14], Theorem 1.3 implies the following improved hardness of approximation results:

Theorem 1.6.

For all ε>0\varepsilon>0, for sufficiently large kk it is NP-hard to approximate the Rooted-kk-Connectivity problem on undirected graphs to within a factor of k1/5εk^{1/5-\varepsilon}, the Vertex-Connectivity Survivable Network Design Problem with connectivity parameters at most kk to within a factor of k1/3εk^{1/3-\varepsilon}, and the Vertex-Connectivity kk-Route Cut Problem to within a factor of k1/3εk^{1/3-\varepsilon}.

We remark that in [CCK08], a weaker form of hardness for the Vertex-Connectivity Survivable Network problem is proved. More precisely, they show an Ω(k1/3/logk)\Omega(k^{1/3}/\log k) integrality gap for the set-pair relaxation of the problem. Our hardness result of k1/3εk^{1/3-\varepsilon} improves upon it, showing that (unless P==NP) no relaxation can yield a better than k1/3εk^{1/3-\varepsilon} factor approximation algorithm.

1.2 Our Techniques

Theorem 1.3 is proved via a composition of an Inner PCP and an Outer PCP. Both of these components incorporate ideas from the proof of the 22-to-11 Games Theorem. The outer PCP is constructed using smooth parallel repetition [KS13, KMS17] while the inner PCP is based on the Grassmann graph [KMS17, DKK+18, DKK+21, KMS23].

The novelty in this current paper, in terms of techniques, is twofold. First, we must consider a Grassmann test in a different regime of parameters (as otherwise we would not be able to get a good alphabet to soundness tradeoff) and in a regime of much lower soundness error. These differences complicate matters considerably. Second, our soundness analysis is more involved than that of the 22-to-11-Games Theorem. As is the case in [KMS17, DKK+18, DKK+21, KMS23], we too use global hyperconractivity, but we do so more extensively. We also require quantitatively stronger versions of global hypercontractivity over the Grasssmann graph which are due to [EKL24]. In addition, our analysis incorporates ideas from the plane versus plane test and direct product testing [RS97, IKW12, MZ23], from classical PCP theory [KS13], as well as from error correcting codes [GRS00]. All of these tools are necessary to prove our main technical statement – Lemma 1.7 below – which is a combinatorial statement that may be of independent interest.

We now elaborate on each one of the components separately.

1.2.1 The Inner PCP

Our Inner PCP is based on the subspace vs subspace low degree test. Below, we first give a general overview of the objective in low-degree testing. We then discuss the traditional notion of soundness as well as a non-traditional notion of soundness for low-degree tests. Finally, we explain the low-degree test used in this paper, the notion of soundness that we need from it, and the way that this notion of soundness is used.

Low degree tests in PCPs.

Low degree tests have been have a vital component in PCPs since their inception, and much attention has been devoted to improving their various parameters. The goal in low-degree testing is to encode a low-degree function f:𝔽qn𝔽qf\colon\mathbb{F}_{q}^{n}\to\mathbb{F}_{q} via a table (or a few tables) of values, in a way that allows for local testing. Traditionally, one picks a parameter \ell\in\mathbb{N} (which is thought of as a constant and is most often just 22) and encodes the function ff by the table TT of restrictions of ff to \ell-dimensional affine subspaces of 𝔽qn\mathbb{F}_{q}^{n}. For the case =2\ell=2, the test associated with this encoding is known as the Plane vs Plane test [RS97]. The Plane vs Plane test proceeds by picking two planes P1P_{1}, P2P_{2} intersecting on a line, and then checking that T[P1]T[P_{1}] and T[P2]T[P_{2}] agree on P1P2P_{1}\cap P_{2}. It is easy to see that the test has perfect completeness, namely that a valid table of restrictions TT passes the test with probability 11. In the other direction, the soundness error of the test – which is a converse type statement – is much less clear (and is crucial towards applications in PCP). In the context of the Plane vs Plane test, it is know that if a table TT, that assigns to each plane a degree dd function, passes the Plane vs Plane test with probability εqc\varepsilon\geqslant q^{-c} (where c>0c>0 is a small absolute constant), then there is a degree dd function ff such that T[P]f|PT[P]\equiv f|_{P} on at least Ω(ε)\Omega(\varepsilon) fraction of the planes.

Nailing down the value of the constant cc for which soundness holds is an interesting open problem which is related to soundness vs alphabet size vs instance size tradeoff in PCPs [MR10, BDN17, MZ23]. Currently, the best known analysis for the Plane vs Plane test [MR10] shows that one may take c=1/8c=1/8. Better analysis is known for higher dimensional encoding [BDN17, MZ23], and for the 33-dimensional version of it a near optimal soundness result is known [MZ23].

Low degree tests in this paper.

In the context of the current paper, we wish to encode linear functions f:𝔽qn𝔽qf\colon\mathbb{F}_{q}^{n}\to\mathbb{F}_{q}, and we do so by the subspaces encoding. Specifically, we set integer parameters 12\ell_{1}\geqslant\ell_{2}, and encode the function ff using the table T1T_{1} of the restrictions of ff to all 1\ell_{1}-dimensional linear subspaces of 𝔽qn\mathbb{F}_{q}^{n}, and the table T2T_{2} of the restrictions of ff to all 2\ell_{2}-dimensional linear subspaces of 𝔽qn\mathbb{F}_{q}^{n}. The test we consider is the natural inclusion test:

  1. 1.

    Sample a random 1\ell_{1}-dimensional subspace L1𝔽qnL_{1}\subseteq\mathbb{F}_{q}^{n} and a random 2\ell_{2}-dimensional subspace L2L1L_{2}\subseteq L_{1}.

  2. 2.

    Read T1[L1]T_{1}[L_{1}], T2[L2]T_{2}[L_{2}] and accept if they agree on L2L_{2}.

As is often the case, the completeness of the test – namely the fact that valid tables T1,T2T_{1},T_{2} pass the test with probability 11 – is clear. The question of most interest then is with regards to the soundness of the test. Namely, what is the smallest ε\varepsilon such that any two tables T1T_{1} and T2T_{2} that assign linear functions to subspaces and pass the test with probability ε\varepsilon, must necessarily “come from” a legitimate linear function ff?

Traditional notion of soundness.

As the alphabet vs soundness tradeoff is key to the discussion herein, we begin by remarking that the alphabet size of the above encoding is q1+q2=Θ(q1)q^{\ell_{1}}+q^{\ell_{2}}=\Theta(q^{\ell_{1}}) (since there are qq^{\ell} distinct linear functions on a linear space of dimension \ell over 𝔽q\mathbb{F}_{q}). Thus, ideally we would like to show that the soundness error of the above test is q(1o(1))1q^{-(1-o(1))\ell_{1}}. Alas, this is false. Indeed, it turns out that one may construct assignments that pass the test with probability at least Ω(max(q2,q21))\Omega(\max(q^{-\ell_{2}},q^{\ell_{2}-\ell_{1}})) that do not have significant correlation with any linear function ff:

  1. 1.

    Taking T1,T2T_{1},T_{2} randomly by assigning to each subspace a random linear function, one can easily see that the test passes with probability Θ(q2)\Theta(q^{-\ell_{2}}).

  2. 2.

    Taking linear subspaces W1,,W100q1𝔽qnW_{1},\ldots,W_{100q^{\ell_{1}}}\subseteq\mathbb{F}_{q}^{n} of co-dimension 11 randomly, and a random linear function fi:Wi𝔽qf_{i}\colon W_{i}\to\mathbb{F}_{q} for each ii, one may choose T1T_{1} and T2T_{2} as follows. For each L1L_{1}, pick a random ii such that L1WiL_{1}\subseteq W_{i} (if such ii exists) and assign T1[L1]=fi|L1T_{1}[L_{1}]=f_{i}|_{L_{1}}. For each L2L_{2}, pick a random ii such that L2WiL_{2}\subseteq W_{i} (if such ii exists) and assign T2[L2]=fi|L2T_{2}[L_{2}]=f_{i}|_{L_{2}}. Taking L2L1L_{2}\subseteq L_{1} randomly, one sees that with constant probability L2L_{2} has Θ(q12)\Theta(q^{\ell_{1}-\ell_{2}}) many possible ii’s, L1L_{1} has Θ(1)\Theta(1) many possible ii’s and furthermore there is at least one ii that is valid for both of them. With probability Ω(q21)\Omega(q^{\ell_{2}-\ell_{1}}) this common ii is chosen for both L1L_{1} and L2L_{2}, and in this case, the test on (L1,L2)(L_{1},L_{2}) passes. It follows that, in expectation, T1,T2T_{1},T_{2} pass the test with probability Ω(q21)\Omega(q^{\ell_{2}-\ell_{1}}).

In light of the above, it makes sense that the best possible alphabet vs soundness tradeoff we may achieve with the subspace encoding is by taking 2=1/2\ell_{2}=\ell_{1}/2. Such a setting of the parameters would give alphabet size R=q1R=q^{\ell_{1}} and (possibly) soundness error Θ(1/R)\Theta(1/\sqrt{R}). There are several issues with this setting however. First, this tradeoff is not good enough for our purposes (which already rules out this setting of parameters). Second, we do not know how to prove that the soundness error of the test is Θ(1/R)\Theta(1/\sqrt{R}) (the best we can do is quadratically off and is Θ(1/R1/4)\Theta(1/R^{1/4})). To address both of these issues, we must venture beyond the traditional notion of soundness.

Non-traditional notion of soundness.

The above test was first considered in the context of the 22-to-11 Games Theorem, wherein one takes q=2q=2 and 2=11\ell_{2}=\ell_{1}-1. In this setting, the test is not sound in the traditional sense; instead, the test is shown to satisfy a non-standard notion of soundness, which nevertheless is sufficient for the purposes of constructing a PCP. More specifically, in [KMS23] it is proved that for all ε>0\varepsilon>0 there is rr\in\mathbb{N} such that for sufficiently large \ell and for tables T1,T2T_{1},T_{2} as above, there are subspaces QW𝔽qnQ\subseteq W\subseteq\mathbb{F}_{q}^{n} with 𝖽𝗂𝗆(Q)+𝖼𝗈𝖽𝗂𝗆(W)r{\sf dim}(Q)+{\sf codim}(W)\leqslant r and a linear function f:W𝔽qf\colon W\to\mathbb{F}_{q} such that

PrQL1W[T1[L1]f|L1]ε(ε)>0.\Pr_{Q\subseteq L_{1}\subseteq W}[T_{1}[L_{1}]\equiv f|_{L_{1}}]\geqslant\varepsilon^{\prime}(\varepsilon)>0.

We refer to the set

{L𝔽qn|𝖽𝗂𝗆(L)=1,QLW}\{L\subseteq\mathbb{F}_{q}^{n}~{}|~{}{\sf dim}(L)=\ell_{1},Q\subseteq L\subseteq W\}

as the zoom in of QQ and zoom out of WW. While this result is good for the purposes of 22-to-11 Games, the dependency between \ell and ε\varepsilon (and thus, between the soundness and the alphabet size) is still not good enough for us.

Our low-degree test.

It turns out that the proper setting of parameters for us is 2=(1δ)1\ell_{2}=(1-\delta)\ell_{1} where δ>0\delta>0 is a small constant. With these parameters, we are able to show that for εq(1δ)1\varepsilon\geqslant q^{-(1-\delta^{\prime})\ell_{1}} (where δ=δ(δ)>0\delta^{\prime}=\delta^{\prime}(\delta)>0 is a vanishing function of δ\delta), if T1T_{1}, T2T_{2} pass the test with probability at least ε\varepsilon, then there are subspaces QWQ\subseteq W with 𝖽𝗂𝗆(Q)+𝖼𝗈𝖽𝗂𝗆(W)r=r(δ){\sf dim}(Q)+{\sf codim}(W)\leqslant r=r(\delta)\in\mathbb{N}, and a linear function f:W𝔽qf\colon W\to\mathbb{F}_{q} such that

PrQL1W[T1[L1]f|L1]ε(ε)=Ω(ε).\Pr_{Q\subseteq L_{1}\subseteq W}[T_{1}[L_{1}]\equiv f|_{L_{1}}]\geqslant\varepsilon^{\prime}(\varepsilon)=\Omega(\varepsilon).

Working in the very small soundness regime of εq(1δ)1\varepsilon\geqslant q^{-(1-\delta^{\prime})\ell_{1}} entails with it many challenges, however. First, dealing with such small soundness requires us to use a strengthening of the global hypercontractivity result of [KMS23] in the form of an optimal level dd inequality due to Evra, Kindler and Lifshitz [EKL24]. Second, in the context of [KMS23], ε\varepsilon^{\prime} could be any function of ε\varepsilon (and indeed it ends up being a polynomial function of ε\varepsilon). In the context of the current paper, it is crucial that ε=ε1+o(1)\varepsilon^{\prime}=\varepsilon^{1+o(1)}, as opposed to, say, ε=ε1.1\varepsilon^{\prime}=\varepsilon^{1.1}. The reason is that, as we are dealing with very small ε\varepsilon, the result would be trivial for ε=ε1.1\varepsilon^{\prime}=\varepsilon^{1.1} and not useful towards the analysis of the PCP (as then ε\varepsilon^{\prime} would be below the threshold q1q^{-\ell_{1}} which represents the agreement a random linear function ff has with T1T_{1}).

1.2.2 Getting List Decoding Bounds

As is usually the case in PCP reductions, we require a list decoding version for our low-degree test. Indeed, using a standard argument we are able to show that in the setting that 2=(1δ)1\ell_{2}=(1-\delta)\ell_{1} and εq(1δ)1\varepsilon\geqslant q^{(1-\delta^{\prime})\ell_{1}}, there is r=r(δ,δ)r=r(\delta,\delta^{\prime})\in\mathbb{N} such that for at least qΘ(1)q^{-\Theta(\ell_{1})} fraction of subspaces Q𝔽qnQ\subseteq\mathbb{F}_{q}^{n} of dimension rr, there exists a subspace WW with co-dimension at most rr and QW𝔽qnQ\subseteq W\subseteq\mathbb{F}_{q}^{n}, as well as a linear function f:W𝔽qf\colon W\to\mathbb{F}_{q}, such that

PrQL1W[T1[L1]f|L1]ε(ε)=Ω(ε).\Pr_{Q\subseteq L_{1}\subseteq W}[T_{1}[L_{1}]\equiv f|_{L_{1}}]\geqslant\varepsilon^{\prime}(\varepsilon)=\Omega(\varepsilon). (1)

This list decoding version theorem alone is not enough. In our PCP construction, we compose the inner PCP with an outer PCP (that we describe below), and analyzing the composition requires decoding global linear functions (from a list decoding version theorem as above) in a coordinated manner between two non communicating parties. Often times, the number of possible global functions that may be decoded is constant, in which case randomly sampling one among them often works. This is not the case for us, though: if (Q,W)(Q,W) and (Q,W)(Q^{\prime},W^{\prime}) are distinct zoom-in and zoom-out pairs for which there are linear functions fQ,Wf_{Q,W} and fQ,Wf_{Q^{\prime},W^{\prime}} satisfying (1), then the functions fQ,Wf_{Q,W} and fQ,Wf_{Q^{\prime},W^{\prime}} could be completely different. Thus, to achieve a coordinated decoding procedure, we must:

  1. 1.

    Facilitate a way for the two parties to agree on a zoom-in and zoom-out pair (Q,W)(Q,W) with noticeable probability.

  2. 2.

    Show that for each (Q,W)(Q,W) there are at most 𝗉𝗈𝗅𝗒(1/ε){\sf poly}(1/\varepsilon) functions fQ,Wf_{Q,W} for which

    PrQL1W[T1[L1]fQ,W|L1]ε.\Pr_{Q\subseteq L_{1}\subseteq W}[T_{1}[L_{1}]\equiv f_{Q,W}|_{L_{1}}]\geqslant\varepsilon^{\prime}.

The second item is precisely the reason we need ε\varepsilon^{\prime} to be ε1+o(1)\varepsilon^{1+o(1)}; any worse dependency, such as ε=ε1.1\varepsilon^{\prime}=\varepsilon^{1.1} would lead to the second item being false. We also remark that the number of functions being 𝗉𝗈𝗅𝗒(1/ε){\sf poly}(1/\varepsilon) is important to us as well. There is some slack in this bound, but a weak quantitative bound such as 𝖾𝗑𝗉(𝖾𝗑𝗉(1/ε)){\sf exp}({\sf exp}(1/\varepsilon)) would have been insufficient for some of our applications. Luckily, such bounds can be deduced from [GRS00] for the case of linear functions.333In the case of higher degree functions (even quadratic functions) some bounds are known [Gop13, BL15] but they would not have been good enough for us.

We now move onto the first item, in which we must facilitate a way for two non-communicating parties to agree on a zoom-in and zoom-out pair (Q,W)(Q,W). It turns out that agreeing on the zoom-in QQ can be delegated to the outer PCP, and we can construct a sound outer PCP game in which the two parties are provided with a coordinated zoom-in QQ. This works because in our list decoding theorem, the fraction of zoom-ins QQ that work is significant. Coordinating zoom-outs is more difficult, and this is where much of the novelty in our analysis lies.

1.2.3 Coordinating Zoom-outs

For the sake of simplicity and to focus on the main ideas, we ignore zoom-ins for now and assume that the list decoding statement holds with no QQ. Thus, the list decoding theorem asserts that there exists a zoom-out WW of constant co-dimension on which there is a global linear function. However, there could be many such zoom-outs WW, say W1,,WmW_{1},\ldots,W_{m} and say all of them were of co-dimension r=Oδ,δ(1)r=O_{\delta,\delta^{\prime}}(1). If the number mm were sufficiently large – say at least q𝗉𝗈𝗅𝗒(1)q^{-{\sf poly}(\ell_{1})} fraction of all co-dimension rr subspaces – then we would have been able to coordinate them in the same way as we coordinate zoom-ins. If the number mm were sufficiently small – say m=q𝗉𝗈𝗅𝗒(1)m=q^{{\sf poly}(\ell_{1})}, then randomly guessing a zoom-out would work well enough. The main issue is that the number mm is intermediate, say m=qnm=q^{\sqrt{n}}.

This issue had already appeared in [KMS17, DKK+18]. Therein, this issue is resolved by showing that if there are at least mq10012m\geqslant q^{100\ell_{1}^{2}} zoom-outs W1,,WmW_{1},\ldots,W_{m} of co-dimension rr, and linear functions f1,,fmf_{1},\ldots,f_{m} on W1,,WmW_{1},\ldots,W_{m} respectively such that

PrLWi[T[L]fi|L]ε\Pr_{L\subseteq W_{i}}[T[L]\equiv f_{i}|_{L}]\geqslant\varepsilon^{\prime}

for all ii, then there exists a zoom out WW of co-dimension strictly less than rr and a linear function f:W𝔽qf\colon W\to\mathbb{F}_{q} such that

PrLW[T[L]f|L]Ω(ε12).\Pr_{L\subseteq W}[T[L]\equiv f|_{L}]\geqslant\Omega(\varepsilon^{\prime 12}).

Thus, if there are too many zoom-outs of a certain co-dimension, then there is necessarily a zoom-out of smaller co-dimension that also works. In that case, the parties could go up to that co-dimension.

This result is not good enough for us, due to the polynomial gap between the agreement between and fif_{i}’s and FF and the agreement between ff an TT. Indeed, in our range of parameters, ε12\varepsilon^{\prime 12} will be below the trivial threshold q1q^{-\ell_{1}} which is the agreement a random linear function ff has with TT, and therefore the promise on the function ff above is meaningless.

We resolve this issue by showing a stronger, essentially optimal version of the above assertion still holds. Formally, we prove:

Lemma 1.7.

For all δ>0\delta>0, rr\in\mathbb{N} there is C>1C>1 such that the following holds for εq(1δ)1\varepsilon^{\prime}\geqslant q^{(1-\delta)\ell_{1}}. Suppose that FF is a table that assigns to each subspace LL of dimension 1\ell_{1} a linear function, and suppose that there are at least mqC1m\geqslant q^{C\ell_{1}} subspaces W1,,WmW_{1},\ldots,W_{m} of co-dimension rr and linear functions fi:Wi𝔽qf_{i}\colon W_{i}\to\mathbb{F}_{q} such that

PrLWi[T[L]fi|L]ε\Pr_{L\subseteq W_{i}}[T[L]\equiv f_{i}|_{L}]\geqslant\varepsilon^{\prime}

for all i=1,,mi=1,\ldots,m. Then, there exists a zoom-out WW of co-dimension strictly smaller than rr and a linear function f:W𝔽qf\colon W\to\mathbb{F}_{q} such that

PrLW[T[L]f|L]Ω(ε).\Pr_{L\subseteq W}[T[L]\equiv f|_{L}]\geqslant\Omega(\varepsilon^{\prime}).

We defer a detailed discussion about Lemma 1.7 and its proof to Section 8, but remark that our proof of Lemma 1.7 is very different from the arguments in [DKK+18] and is significantly more involved. Our proof uses tools from [KMS17, DKK+18], tools from the analysis of the classical Plane vs Plane and direct product testing [RS97, IKW12, MZ23], global hypercontractivity [EKL24] as well as Fourier analysis over the Grassmann graph.

1.2.4 The Outer PCP

Our outer PCP game is the outer PCP of [KMS17, DKK+18], which is a smooth parallel repetition of the equation versus variables game of Hastad [Hås01] (or of [KP06] for the application to Quadratic Programming). As in there, we equip this game with the “advice” feature to facilitate zoom-in coordination (as discussed above). For the sake of completeness we elaborate on the construction of the outer PCP below.

We start with an instance of 33-Lin that has a large gap between the soundness and completeness. Namely, we start with an instance (X,E)(X,E) of linear equations over 𝔽q\mathbb{F}_{q} in which each equation has the form axi1+bxi2+cxi3=dax_{i_{1}}+bx_{i_{2}}+cx_{i_{3}}=d. It is known [Hås01] that for all η>0\eta>0, it is NP-hard to distinguish between the following two cases:

  1. 1.

    YES case: 𝗏𝖺𝗅(X,E)1η{\sf val}(X,E)\geqslant 1-\eta.

  2. 2.

    NO case: 𝗏𝖺𝗅(X,E)1.1q{\sf val}(X,E)\leqslant\frac{1.1}{q}.

Given the instance (X,E)(X,E), we construct a 22-Prover-11-Round game, known as the smooth equation versus variable game with rr-advice as follows. The verifier has a smoothness parameter β>0\beta>0 and picks a random equation ee, say axi1+bxi2+cxi3=dax_{i_{1}}+bx_{i_{2}}+cx_{i_{3}}=d, from (X,E)(X,E). Then:

  1. 1.

    With probability 1β1-\beta the verifier takes U=V={xi1,xi2,xi3}U=V=\{x_{i_{1}},x_{i_{2}},x_{i_{3}}\} and vectors u1=v1,,ur=vr𝔽qUu_{1}=v_{1},\ldots,u_{r}=v_{r}\in\mathbb{F}_{q}^{U} sampled uniformly and independently.

  2. 2.

    With probability β\beta, the verifier sets U={xi1,xi2,xi3}U=\{x_{i_{1}},x_{i_{2}},x_{i_{3}}\}, chooses a set consisting of a single variable VUV\subseteq U uniformly at random. The verifier picks v1,,vr𝔽qVv_{1},\ldots,v_{r}\in\mathbb{F}_{q}^{V} uniformly and independently and appends to each viv_{i} the value 0 in the coordinates of UVU\setminus V to get u1,,uru_{1},\ldots,u_{r}.

After that, the verifier sends UU and u1,,uru_{1},\ldots,u_{r} to the first prover and VV and v1,,vrv_{1},\ldots,v_{r} to the second prover. The verifier expects to get from them 𝔽q\mathbb{F}_{q} assignments to the variables in UU and in VV, and accepts if and only if these assignments are consistent, and furthermore the assignment to UU satisfies the equation ee.

Denoting the equation versus variable game by Ψ{\sf\Psi}, it is easy to see that if 𝗏𝖺𝗅(X,E)1η{\sf val}(X,E)\geqslant 1-\eta, then 𝗏𝖺𝗅(Ψ)1η{\sf val}(\Psi)\geqslant 1-\eta, and if 𝗏𝖺𝗅(X,E)1.1/q{\sf val}(X,E)\leqslant 1.1/q, then 𝗏𝖺𝗅(Ψ)1Ω(qrβ){\sf val}(\Psi)\leqslant 1-\Omega(q^{-r}\beta). The gap between 1η1-\eta and 1Ω(qrβ)1-\Omega(q^{-r}\beta) is too weak for us, and thus we apply parallel repetition.

In the parallel repetition of the smooth equation versus variable game with advice, denoted by Ψk\Psi^{\otimes k}, the verifier picks kk equations uniformly and independently e1,,eke_{1},\ldots,e_{k}, and picks UiU_{i}, u1,i,,ur,iu_{1,i},\ldots,u_{r,i} and ViV_{i}, v1,i,,vr,iv_{1,i},\ldots,v_{r,i} for each i=1,,ki=1,\ldots,k from eie_{i} independently. Thus, the questions of the provers may be seen as U=U1UkU=U_{1}\cup\ldots\cup U_{k} and V=V1VkV=V_{1}\cup\ldots\cup V_{k} and their advice is uj=(uj,1,,uj,k)𝔽qU\vec{u}_{j}=(u_{j,1},\ldots,u_{j,k})\in\mathbb{F}_{q}^{U} for j=1,,rj=1,\ldots,r and vj=(vj,1,,vj,k)𝔽qV\vec{v}_{j}=(v_{j,1},\ldots,v_{j,k})\in\mathbb{F}_{q}^{V} for j=1,,rj=1,\ldots,r respectively. The verifier expects to get from the first prover a vector in 𝔽qU\mathbb{F}_{q}^{U} which specifies an 𝔽q\mathbb{F}_{q} assignment to UU, and from the second prover a vector in 𝔽qV\mathbb{F}_{q}^{V} specifying an 𝔽q\mathbb{F}_{q} assignment to VV. The verifier accepts if and only if these assignments are consistent and the assignment of the first prover satisfies all of e1,,eke_{1},\ldots,e_{k}. It is clear that if 𝗏𝖺𝗅(X,E)1η{\sf val}(X,E)\geqslant 1-\eta, then 𝗏𝖺𝗅(Ψm)1kη{\sf val}(\Psi^{\otimes m})\geqslant 1-k\eta. Using the parallel repetition theorem of Rao [Rao11] (albeit not in a completely trivial way) we argue that if 𝗏𝖺𝗅(X,E)1.1q{\sf val}(X,E)\leqslant\frac{1.1}{q}, then 𝗏𝖺𝗅(Ψk)2Ω(βqrk){\sf val}(\Psi^{\otimes k})\leqslant 2^{-\Omega(\beta q^{-r}k)}. The game Ψk\Psi^{\otimes k} is our outer PCP game.

Remark 1.8.

We remark that in the case of the Quadratic Programming application, we require a hardness result in which the completeness is very close to 11 in the form of Theorem 2.1. The differences between the reduction in that case and the reduction presented above are mostly minor, and amount to picking the parameters a bit differently. There is one significant difference in the analysis; we require a much sharper form of the “covering property” used in [KMS17, DKK+18], as elaborated on in Section 1.2.6

1.2.5 Composing the Outer PCP and the Inner PCP Game

To compose the outer and inner PCPs, we take the outer PCP game, only keep the questions UU to the first prover and consider an induced 22-Prover-11-Round game on it. The alphabet is 𝔽q3k\mathbb{F}_{q}^{3k}, that given a question UU specifies an 𝔽q\mathbb{F}_{q} assignment to the variables of UU. There is a constraint between UU and UU^{\prime} if there is a question VV to the second prover such that VUUV\subseteq U\cap U^{\prime}. Denoting the assignments to UU and UU^{\prime} by sUs_{U} and sUs_{U^{\prime}}, the constraint between UU and UU^{\prime} is that sUs_{U} satisfies all of the equations that form UU, sUs_{U^{\prime}} satisfies all of the equations that form UU^{\prime}, and sUs_{U}, sUs_{U^{\prime}} agree on UUU\cap U^{\prime}.

The composition amounts to replacing each question UU with a copy of our inner PCP. Namely, we identify between the question UU and the space 𝔽qU\mathbb{F}_{q}^{U}, and then replace UU by a copy of the 2,1\ell_{2},\ell_{1} sub-spaces graph of 𝔽qU\mathbb{F}_{q}^{U}. The answer sUs_{U} is naturally identified with the linear function fU(x)=sU,xf_{U}(x)=\langle s_{U},x\rangle, which is then encoded by the sub-spaces encoding via tables of assignments T1,UT_{1,U} and T2,UT_{2,U}.

The constraints of the composed PCP must check that: (1) side conditions: the encoded vector sUs_{U} satisfies the equations of UU, and (2) consistency: sUs_{U} and sUs_{U^{\prime}} agree on UUU\cap U^{\prime}.

The first set of constraints is addressed by the folding technique, which we omit from this discussion. The second set of constraints is addressed by the 1\ell_{1} vs 2\ell_{2} subspace test, except that we have to modify it so that it works across blocks UU and UU^{\prime}. This completes the description of the composition step of the other PCP and the inner PCP, and thereby the description of our reduction.

1.2.6 The Covering Property

We end this introductory section by discussing the covering property. The covering property is an important feature of our outer PCP construction which enables the composition step to go through. The covering property first appeared in [KS13] and later more extensively in the context of the 22-to-11 Games [KMS17, DKK+18]. To discuss the covering property, let kk\in\mathbb{N} be thought of as large, let β(0,1)\beta\in(0,1) be thought of as k0.99k^{-0.99} and consider sets U1,,UkU_{1},\ldots,U_{k} consisting of distinct element, each UiU_{i} has size 33 (in our context, UiU_{i} will be the set of variables in the iith equation the verifier chose). Let U=U1UkU=U_{1}\cup\ldots\cup U_{k}, and consider the following two distributions over tuples in 𝔽qU\mathbb{F}_{q}^{U}:

  1. 1.

    Sample x1,,x𝔽qUx_{1},\ldots,x_{\ell}\in\mathbb{F}_{q}^{U} uniformly.

  2. 2.

    For each ii independently, take Vi=UiV_{i}=U_{i} with probability 1β1-\beta and otherwise take ViUiV_{i}\subseteq U_{i} randomly of size 11, then set V=V1VkV=V_{1}\cup\ldots\cup V_{k}. Sample x1,,x𝔽qVx_{1},\ldots,x_{\ell}\in\mathbb{F}_{q}^{V} uniformly and lift them to points in 𝔽qU\mathbb{F}_{q}^{U} by appending 0’s in UVU\setminus V. Output the lifted points.

In [KMS17] it is shown that the two distributions above are q3βkq^{3\ell}\beta\sqrt{k} close in statistical distance, which is good enough for the purposes of Theorem 1.3. However, this is not good enough for Theorem 1.4. 444 The reason is that letting NN be the size of the instance we produce, it holds that kk is roughly logarithmic logN\log N and qq^{\ell} is the alphabet size. To have small statistical distance, we must have kq6k\leqslant q^{6\ell}, hence the soundness error could not go lower than (logN)1/6(\log N)^{-1/6}. Carrying out a different analysis, we are able to show that the two distributions are close with better parameters and in a stronger sense: there exists a set EE of \ell tuples which has negligible measure in both distributions, such that each tuple not in EE is assigned the same probability under the two distribution up to factor (1+o(1))(1+o(1)). We are able to prove this statement provided that kk is only slightly larger than q2q^{2\ell}.

The issue with the above two distributions is that they are actually far from each other if, say, k=q1.9k=q^{1.9\ell}. To see that, one can notice that the expected number of ii’s such that each one of x1,,xx_{1},\ldots,x_{\ell} has the form (a,0,0)𝔽q3(a,0,0)\in\mathbb{F}_{q}^{3} on coordinates corresponding to UiU_{i} is very different. In the first distribution, this expectation is Θ(q2k)\Theta(q^{-2\ell}k) which is less than 1, whereas in the second distribution it is at least βkk0.01\beta k\geqslant k^{0.01}.

To resolve this issue and to go all the way through in the Quadratic Programming application, we have to modify the distributions in the covering property so that (a) they will be close even if k=q1.01k=q^{1.01\ell}, and (b) we can still use these distributions in the composition step in our analysis of the PCP construction. Indeed, this is the route we take, and the two distributions we use are defined as follows:

  1. 1.

    Sample x1,,x𝔽qUx_{1},\ldots,x_{\ell}\in\mathbb{F}_{q}^{U} uniformly.

  2. 2.

    For each ii independently, take Vi=UiV_{i}=U_{i} with probability 1β1-\beta and otherwise take ViUiV_{i}\subseteq U_{i} randomly of size 11, then set V=V1VkV=V_{1}\cup\ldots\cup V_{k}. Sample x1,,x𝔽qVx_{1},\ldots,x_{\ell}\in\mathbb{F}_{q}^{V} uniformly, and let wi=1Ui𝔽qUw_{i}=1_{U_{i}}\in\mathbb{F}_{q}^{U} be the vector that has 11 on coordinates of UiU_{i} and 0 everywhere else. Lift the points x1,,xx_{1},\ldots,x_{\ell} to x1,,x𝔽qUx_{1}^{\prime},\ldots,x_{\ell}^{\prime}\in\mathbb{F}_{q}^{U} by appending 0’s in UVU\setminus V and take yj=xj+i=1kαi,jwiy_{j}=x_{j}+\sum\limits_{i=1}^{k}\alpha_{i,j}w_{i} where αi,j\alpha_{i,j} are independent random elements from 𝔽q\mathbb{F}_{q}. Output y1,,yy_{1},\ldots,y_{\ell}.

We show that for a suitable choice of kk and β\beta, these distributions are close even in the case that k=q1.01k=q^{1.01\ell}. 555More speifically, one takes a small c>0c>0 and chooses β=k2c/31\beta=k^{2c/3-1}, k=q(1+c)k=q^{(1+c)\ell}. Indeed, as a sanity check one could count the expected number of appearances of blocks of the form (0,a,0)𝔽3(0,a,0)\in\mathbb{F}^{3} and see they are very close (q2kq^{-2\ell}k versus (1β)q2k+βkq(1-\beta)q^{-2\ell}k+\beta kq^{-\ell}). In this setting of parameters, kk is roughly equal to the alphabet size – which can be made to be equal (logN)1o(1)(\log N)^{1-o(1)} under quasi-polynomial time reductions – it is sufficient to get the result of Theorem 1.4.

Remark 1.9.

We remark that a tight covering property is crucial for obtaining the tight hardness of approximation factor in Theorem 1.4. In the reduction from 22-Prover-11-Round games to Quadratic Programs, which is due to [ABH+05], the size of the resulting instance is exponential in the alphabet size and the soundness error remains roughly the same. In our case the alphabet size is roughly kk hence the instance size is dominated by N=2Θ(k1+o(1))N=2^{\Theta(k^{1+o(1)})}. If our analysis required k=qCk=q^{C\ell}, then even showing an optimal soundness of q(1o(1))q^{-(1-o(1))\ell} for the 22-Prover-11-Round game would only yield a factor of (logN)1/Co(1)(\log N)^{1/C-o(1)} hardness for quadratic programming.

2 Preliminaries

2.1 The Grassmann Graph

In this section we present the Grassmann graph and some Fourier analytic tools on it that are required for our analysis of the inner PCP. Throughout this section, we fix parameters n,n,\ell with 1n1\ll\ell\ll n, and a prime power qq.

2.1.1 Basic Definitions

The Grassmann graph 𝖦𝗋𝖺𝗌𝗌q(n,){\sf Grass}_{q}(n,\ell) is defined as follows.

  • The vertex set corresponds to the set of \ell-dimensional subspaces L𝔽qnL\subseteq\mathbb{F}_{q}^{n}.

  • The edge set corresponds to all pairs (L,L)(L,L^{\prime}) of \ell-dimensional subspaces L,L𝔽qnL,L^{\prime}\subseteq\mathbb{F}_{q}^{n} such that dim(LL)=1\dim(L\cap L^{\prime})=\ell-1.

At times we will have a vector space VV over 𝔽q\mathbb{F}_{q}, and thus we may identify VV with 𝔽qn\mathbb{F}_{q}^{n} and work with the Grassmann graph on the \ell-dimensional subspaces LVL\subseteq V. We may also use 𝖦𝗋𝖺𝗌𝗌q(V,){\sf Grass}_{q}(V,\ell) to denote this graph, which is isomporphic to 𝖦𝗋𝖺𝗌𝗌q(n,){\sf Grass}_{q}(n,\ell) if dim(V)=n\dim(V)=n. Abusing notation, we also use 𝖦𝗋𝖺𝗌𝗌q(n,){\sf Grass}_{q}(n,\ell) to denote the set of all \ell-dimensional subspaces in 𝔽qn\mathbb{F}_{q}^{n}. Throughout, we denote by L2(𝖦𝗋𝖺𝗌𝗌q(n,))L_{2}({\sf Grass}_{q}(n,\ell)) the set of complex valued functions F:[𝔽qn]qF\colon\begin{bmatrix}{\mathbb{F}_{q}^{n}}\\ {\ell}\end{bmatrix}_{q}\to\mathbb{C}.

The number of \ell-dimensional subspaces of 𝔽qn\mathbb{F}_{q}^{n} is counted by the Gaussian binomial coefficient, [n]q\begin{bmatrix}{n}\\ {\ell}\end{bmatrix}_{q}. The following standard fact gives a formula for the Gaussian binomial coefficients, and we omit the proof.

Fact 2.1.

Suppose 1n21\leqslant\ell\leqslant\frac{n}{2}, then the number of vertices in 𝖦𝗋𝖺𝗌𝗌q(n,){\sf Grass}_{q}(n,\ell) is given by

[n]q=i=01qnqiqqi.\begin{bmatrix}{n}\\ {\ell}\end{bmatrix}_{q}=\prod_{i=0}^{\ell-1}\frac{q^{n}-q^{i}}{q^{\ell}-q^{i}}.

Abusing notations, we denote by [V]q\begin{bmatrix}{V}\\ {\ell}\end{bmatrix}_{q} the set of \ell dimensional subspaces of VV.

Zoom ins and Zoom outs.

A feature of the Grassmann graph is that it contains many copies of lower dimensional Grassmann graphs as induced subgraphs. These subgraphs are precisely the zoom-ins and and zoom-outs referred to in the introduction, and they play a large part in the analysis of our inner PCP and final PCP. For subspaces QW𝔽qnQ\subseteq W\subseteq\mathbb{F}_{q}^{n}, let

𝖹𝗈𝗈𝗆[Q,W]={L𝖦𝗋𝖺𝗌𝗌q(n,)|QLW}.{\sf Zoom}[Q,W]=\{L\in{\sf Grass}_{q}(n,\ell)\;|\;Q\subseteq L\subseteq W\}.

We refer to QQ as a zoom-in and WW as a zoom-out. When W=𝔽qnW=\mathbb{F}_{q}^{n}, 𝖹𝗈𝗈𝗆[Q,W]{\sf Zoom}[Q,W] is the zoom-in on QQ, and when Q={0}Q=\{0\}, 𝖹𝗈𝗈𝗆[Q,W]{\sf Zoom}[Q,W] is the zoom-out on WW.

2.1.2 Pseudo-randomness over the Grassmann graph

One notion that will be important to us is (r,ε)(r,\varepsilon)-pseudo-randomness, which measures how much FF can deviate from its expectation on a zoom-in/zoom-out restrictions of “size rr”. For all of our applications, FF and GG will both be indicator functions of some sets of vertices, so it will be helpful to think of this case for the remainder of the section. 666We remark that the results we state have more general versions that apply to wider classes of functions. We refrain from stating them in this generality for sake of simplicity. Let μ(F)=𝔼L𝖦𝗋𝖺𝗌𝗌q(n,)[F(L)]\mu(F)=\mathop{\mathbb{E}}_{L\in{\sf Grass}_{q}(n,\ell)}[F(L)] (for indicator functions, this is simply the measure of the indicated set). For subspaces QW𝔽qnQ\subseteq W\subseteq\mathbb{F}_{q}^{n}, define

μQ,W(F)=𝔼L𝖦𝗋𝖺𝗌𝗌q(n,)[F(L)|QLW].\mu_{Q,W}(F)=\mathop{\mathbb{E}}_{L\in{\sf Grass}_{q}(n,\ell)}[F(L)\;|\;Q\subseteq L\subseteq W].
Definition 2.2.

We say that a Boolean function F:G(n,){0,1}F\colon G(n,\ell)\to\{0,1\} is (r,ε)(r,\varepsilon)-pseudo-random if for all QW𝔽qnQ\subseteq W\subseteq\mathbb{F}_{q}^{n} satisfying dim(Q)+codim(W)=r\dim(Q)+\operatorname{codim}(W)=r, we have

μQ,W(F)ε.\mu_{Q,W}(F)\leqslant\varepsilon.

We will often say that a set S𝖦𝗋𝖺𝗌𝗌q(n,)S\subseteq{\sf Grass}_{q}(n,\ell) is (r,ε)(r,\varepsilon)-pseudo-random if its indicator function is. Because the Grassmann graph is not a small-set expanders, there are small sets in it that do not look “random” with respect to some combinatorial counting measures (such as edges between sets, expansion and so on). Intuitively, a small set SS which is highly pseudo-random will exhibit random-like structure with respect to several combinatorial measures of interest, and the two lemmas below are instantiations of it required in our proof. The proof proceed by reducing them to similar statements about the Bi-linear scheme, which can then be proved directed by appealing to global hypercontractivity results of [EKL23, EKL24].

For the analysis of the inner PCP, we require the following lemma, which bounds the number of edges between 𝖦𝗋𝖺𝗌𝗌q(n,2)\mathcal{L}\subseteq{\sf Grass}_{q}(n,2\ell) and 𝖦𝗋𝖺𝗌𝗌q(n,2(1δ)){\sf Grass}_{q}(n,2(1-\delta)\ell) when \mathcal{L} is (r,ε)(r,\varepsilon)-pseudo-random.

Lemma 2.3.

Let F:𝖦𝗋𝖺𝗌𝗌q(n,2){0,1}F\colon\operatorname{{\sf Grass}}_{q}(n,2\ell)\to\{0,1\} and G:𝖦𝗋𝖺𝗌𝗌q(n,2(1δ)){0,1}G\colon\operatorname{{\sf Grass}}_{q}(n,2(1-\delta)\ell)\to\{0,1\} be Boolean functions such that 𝔼L[F(L)]=α\mathop{\mathbb{E}}_{L}[F(L)]=\alpha, 𝔼R[G(R)]=β\mathop{\mathbb{E}}_{R}[G(R)]=\beta, and suppose that FF is (r,ε)(r,\varepsilon) pseudo-random. Then for all t4t\geqslant 4 that are powers of 22,

𝒯F,GqOt,r(1)β(t1)/tε2t/(2t1)+qrδαβ.\langle\mathcal{T}F,G\rangle\leqslant q^{O_{t,r}(1)}\beta^{(t-1)/t}\varepsilon^{2t/(2t-1)}+q^{-r\delta\ell}\sqrt{\alpha\beta}.
Proof.

Deferred to Section A

We also need the following lemma, asserting that if a not-too-small set SS is highly pseudo-random, then its density remains nearly the same on all zoom-ins.

Lemma 2.4.

For all ξ>0\xi>0, the following holds for sufficiently large \ell. Suppose that ξ3\ell^{\prime}\geqslant\frac{\xi}{3}\ell, δ2=ξ100\delta_{2}=\frac{\xi}{100}, and let VV^{\star} be a subspace such that dim(V)2\dim(V^{\star})\geqslant\ell^{\prime 2}. Let 𝖦𝗋𝖺𝗌𝗌q(V,)\mathcal{L}^{\star}\subseteq{\sf Grass}_{q}(V^{\star},\ell^{\prime}) have measure μ()=ηq2\mu(\mathcal{L}^{\star})=\eta\geqslant q^{-2\ell} and set Z={zV||μz()η|η10}Z=\{z\in V^{\star}\;|\;|\mu_{z}(\mathcal{L}^{\star})-\eta|\leqslant\frac{\eta}{10}\}. If \mathcal{L}^{\star} is (1,qδ2/100)(1,q^{\delta_{2}\ell/100})-pseudo-random, then

|Z|(1q2)|V|.|Z|\geqslant\left(1-q^{\frac{\ell^{\prime}}{2}}\right)|V^{\star}|.
Proof.

The proof is deferred to Appendix A.4

At times we will also use the term global to refer to sets whose indicator functions are pseudo-random. That is, we say that a set is (r,ε)(r,\varepsilon)-global if its indicator is (r,ε)(r,\varepsilon)-pseudo-random.

2.2 Hardness of 3LIN

In this section we cite several hardness of approximation results for the problem of solving linear equations over finite fields, which are the starting point of our reduction. We begin by defining the 𝟥𝖫𝗂𝗇{\sf 3Lin} and the 𝖦𝖺𝗉𝟥𝖫𝗂𝗇{\sf Gap3Lin} problem.

Definition 2.5.

For a prime power qq, an instance of 𝟥𝖫𝗂𝗇{\sf 3Lin} is (X,𝖤𝗊)(X,{\sf Eq}) which consists of a set of variables XX and a set of linear equations 𝖤𝗊{\sf Eq} over 𝔽q\mathbb{F}_{q}. Each equation in 𝖤𝗊{\sf Eq} depends on exactly three variables in XX, each variable appears in at most 1010 equations, and any two distinct equations in 𝖤𝗊{\sf Eq} share at most a single variable.

The goal in the 𝟥𝖫𝗂𝗇{\sf 3Lin} problem is to find an assignment A:X𝔽qA\colon X\to\mathbb{F}_{q} satisfying as many of the equations in EE as possible. The maximum fraction of equations that can be satisfied is called the value of the instance. We remark that usually in the literature, the condition that two equations in EE share at most a single variable is not included in the definition of 𝟥𝖫𝗂𝗇{\sf 3Lin}, as well the the bound on the number of occurences of each variable.

For 0<s<c10<s<c\leqslant 1, the problem 𝖦𝖺𝗉𝟥𝖫𝗂𝗇[c,s]{\sf Gap3Lin}[c,s] is the promise problem wherein the input is an instance (X,E)(X,E) of 𝟥𝖫𝗂𝗇{\sf 3Lin} promised to either have value at least cc or at most ss, and the goal is to distinguish between these two cases. The problem 𝖦𝖺𝗉𝟥𝖫𝗂𝗇[c,s]{\sf Gap3Lin}[c,s] with various settings of cc and ss will be the starting point for our reductions.

To prove Theorem 1.3, we shall use the classical result of Håstad [Hås01]. This result says that for general 𝟥𝖫𝗂𝗇{\sf 3Lin} instances (i.e., without the additional condition that two equations share at most a single variable), the problem 𝖦𝖺𝗉𝟥𝖫𝗂𝗇[1ε,1/q+ε]{\sf Gap3Lin}[1-\varepsilon,1/q+\varepsilon] is NP-hard for all constant qq\in\mathbb{N} and ε>0\varepsilon>0. This result implies the following theorem by elementary reductions:

Theorem 2.1.

There exists s<1s<1 such that for every constant η>0\eta>0 and prime qq, 𝖦𝖺𝗉𝟥𝖫𝗂𝗇[1η,s]{\sf Gap3Lin}\left[1-\eta,s\right] is 𝖭𝖯{\sf NP}-hard.

To prove Theorem 1.4 we will need a hardness result for 𝟥𝖫𝗂𝗇{\sf 3Lin} with completeness close to 11, and we will use a hardness result of Khot and Ponnuswami [KP06]. Once again, their result does not immediately guarantee the fact that any two equations share at most a single variable, however once again this property may be achieved by an elementary reduction.

Theorem 2.2.

There is a reduction from 𝖲𝖠𝖳{\sf SAT} with size nn to a 𝖦𝖺𝗉𝟥𝖫𝗂𝗇[1η,1ε]{\sf Gap3Lin}[1-\eta,1-\varepsilon] instance with size NN over a field 𝔽q\mathbb{F}_{q} of characteristic 22, where,

  • Both NN and the running time of the reduction are bounded by 2O(log2n)2^{O(\log^{2}n)}

  • η2Ω(logN)\eta\leqslant 2^{-\Omega(\sqrt{\log N})}.

  • εΩ(1log3N)\varepsilon\geqslant\Omega\left(\frac{1}{\log^{3}N}\right).

3 The Outer PCP

In this section, we describe our outer PCP game. In short, our outer PCP is a smooth parallel repetition of the variable versus equation game with advice. This outer PCP was first considered in [KS13] without the advice feature, and then in [KMS17] with the advice feature.

3.1 The Outer PCP construction

Let ε1<ε2\varepsilon_{1}<\varepsilon_{2} be parameters that determine the completeness and the soundness our 𝖦𝖺𝗉𝟥𝖫𝗂𝗇{\sf Gap3Lin}. Our reduction starts with the 𝖦𝖺𝗉𝟥𝖫𝗂𝗇[1ε1,1ε2]{\sf Gap3Lin}[1-\varepsilon_{1},1-\varepsilon_{2}] problem, and we fix an instance of it (X,E)(X,E) for the rest of this section. Our presentation is gradual, and we begin by presenting the basic Variable versus Equation Game. We then equip it with the additional features of smoothness and advice.

3.1.1 The Variable versus Equation Game

We first convert the instance (X,E)(X,E) into an instance of 22-Prover-11-Round Games, and it will be convenient for us to describe it in the active view with a verifier and 22 provers.

In the Variable versus Equation game, the verifier picks an equation eEe\in E uniformly at random, and then chooses a random variable xex\in e. The verifier sends the question ee, i.e. the three variables appearing in ee, to the first prover, and sends the variable xx to the second prover. The provers are expected to answer with assignments to their received variables, and the verifier accepts if and only if the two assignments agree on xx and the first prover’s assignment satisfies the equation ee. If the verifier accepts then we also say that the provers pass. This game has the following completeness and soundness, which are both easy to see (we omit the formal proof):

  1. 1.

    Completeness: If (X,E)(X,E) has an assignment satisfying 1ε1-\varepsilon-fraction of the equations, then the prover’s have a strategy that passes with probability at least 1ε1-\varepsilon.

  2. 2.

    Soundness: If (X,𝖤𝗊)(X,{\sf Eq}) has no assignment satisfying more than 1ε1-\varepsilon-fraction of the equations, then the prover’s can pass with probability at most 1ε31-\frac{\varepsilon}{3}.

3.1.2 The Smooth Equation versus Variable Game

We next describe a smooth version of the Variable versus Equation game. In this game, the verifier has a parameter β(0,1]\beta\in(0,1], and it proceeds as follows:

  1. 1.

    The verifier chooses an equation eEe\in E uniformly, and lets UU be the set of variables in ee.

  2. 2.

    With probability 1β1-\beta, the prover chooses V=UV=U. With probability β\beta, the prover chooses VUV\subseteq U randomly of size 11.

  3. 3.

    The verifier sends UU to the first prover, and VV to the second prover.

  4. 4.

    The provers respond with assignments to the variables they receive, and the verifier accepts if and only if their assignments agree on VV and the assignment to UU satisfies the equation ee.

The smooth Variable versus Equation game has the following completeness and soundness property, which are again easily seen to hold (we omit the formal proof).

  1. 1.

    Completeness: If (X,E)(X,E) has an assignment satisfying 1ε1-\varepsilon fraction of the equations, then the provers have a strategy that passes with probability at least 1ε1-\varepsilon.

  2. 2.

    Soundness: If (X,𝖤𝗊)(X,{\sf Eq}) has no assignment satisfying more than 1ε1-\varepsilon fraction of the equations, then the provers can pass with probability at most 1βε31-\frac{\beta\varepsilon}{3}.

3.1.3 The Smooth Equation versus Variable Game with Advice

Next, we introduce the feature of advice into the smooth Variable versus Equation Game. This “advice” acts as shared randomness which may help the provers in their strategy; we show though that it does not considerably change the soundness. The game is denoted by Gβ,rG_{\beta,r} for β(0,1]\beta\in(0,1] and rr\in\mathbb{N}, and proceeds as follows:

  1. 1.

    The verifier chooses an equation eEe\in E uniformly, and lets UU be the set of variables in ee.

  2. 2.

    With probability 1β1-\beta, the verifier chooses V=UV=U. With probability β\beta, the verifier chooses VUV\subseteq U randomly of size 11.

  3. 3.

    The verifier picks vectors v1,,vr𝔽qVv_{1},\ldots,v_{r}\in\mathbb{F}_{q}^{V} uniformly and independently. If U=VU=V the verifier takes ui=viu_{i}=v_{i} for all ii, and otherwise the verifier takes the vectors u1,,ur𝔽qUu_{1},\ldots,u_{r}\in\mathbb{F}_{q}^{U} where for all i=1,,ri=1,\ldots,r, the vector uiu_{i} agrees with viv_{i} on the coordinate of VV, and is 0 in the coordinates of UVU\setminus V.

  4. 4.

    The verifier sends UU and u1,,uru_{1},\ldots,u_{r} to the first prover, and VV and v1,,vrv_{1},\ldots,v_{r} to the second prover.

  5. 5.

    The provers respond with assignments to the variables they receive, and the verifier accepts if and only if their assignments agree on VV and the assignment to UU satisfies the equation ee.

Below we state the completeness and soundness of this game:

  1. 1.

    Completeness: If (X,E)(X,E) has an assignment satisfying 1ε1-\varepsilon fraction of the equations, then the provers have a strategy that passes with probability at least 1ε1-\varepsilon. This is easy to see.

  2. 2.

    Soundness: If (X,𝖤𝗊)(X,{\sf Eq}) has no assignment satisfying more than 1ε1-\varepsilon fraction of the equations, then the provers can pass with probability at most 1qrβε31-\frac{q^{-r}\beta\varepsilon}{3}. Indeed, suppose that the provers can win the game with probability at least 1η1-\eta. Note that with probability at least βqr\beta q^{-r} it holds that UVU\neq V and all the vectors u1,,uru_{1},\ldots,u_{r} and v1,,vrv_{1},\ldots,v_{r} are all 0, in which case the provers play the standard equation versus variable game. Thus, the provers’ strategy wins in the latter game with probability at least 1ηqrβ>1ε31-\frac{\eta}{q^{-r}\beta}>1-\frac{\varepsilon}{3}, and contradiction.

3.1.4 Parallel Repetition of the Smooth Equation versus Variable Game with Advice

Finally, our Outer PCP is then the kk-fold parallel repetition of Gβ,rG_{\beta,r}, which we denote by Gβ,rkG^{\otimes k}_{\beta,r}. Below is a full description of it:

  1. 1.

    The verifier chooses equations e1,,ekEe_{1},\ldots,e_{k}\in E uniformly and independently, and lets UiU_{i} be the set of variables in eie_{i}.

  2. 2.

    For each ii independently, with probability 1β1-\beta, the verifier chooses Vi=UiV_{i}=U_{i}. With probability β\beta, the verifier chooses ViUiV_{i}\subseteq U_{i} randomly of size 11.

  3. 3.

    For each ii independently, the verifier picks a vectors v1i,,vri𝔽qVv_{1}^{i},\ldots,v_{r}^{i}\in\mathbb{F}_{q}^{V} uniformly and independently. If Ui=ViU_{i}=V_{i} the verifier takes uji=vjiu_{j}^{i}=v_{j}^{i} for j=1,,rj=1,\ldots,r, and otherwise the verifier takes the vectors u1i,,uri𝔽qUu_{1}^{i},\ldots,u_{r}^{i}\in\mathbb{F}_{q}^{U} where for all j=1,,rj=1,\ldots,r, the vector ujiu_{j}^{i} agrees with vjiv_{j}^{i} on the coordinate of ViV_{i}, and is 0 in the coordinates of UiViU_{i}\setminus V_{i}.

  4. 4.

    The verifier sets U=i=1kUiU=\bigcup_{i=1}^{k}U_{i} and uj=(uj1,,ujk)u_{j}=(u_{j}^{1},\ldots,u_{j}^{k}) for each j=1,rj=1,\ldots r, and V=i=1kViV=\cup_{i=1}^{k}V_{i} and vj=(vj1,,vjk)v_{j}=(v_{j}^{1},\ldots,v_{j}^{k}) for each j=1,,rj=1,\ldots,r. The verifier sends UU and u1,,uru_{1},\ldots,u_{r} to the first prover, and VV and v1,,vrv_{1},\ldots,v_{r} to the second prover.

  5. 5.

    The provers respond with assignments to the variables they receive, and the verifier accepts if and only if their assignments agree on VV and the assignment to UU satisfies the equations e1,,eke_{1},\ldots,e_{k}.

Next, we state the completeness and the soundness of the game Gβ,rkG_{\beta,r}^{\otimes k}, and we begin with its completeness.

Claim 3.1.

If (X,E)(X,E) has an assignment satisfying at least 1ε1-\varepsilon of the equations, then the provers can win Gβ,rkG_{\beta,r}^{\otimes k} with probability at least 1kε1-k\varepsilon.

Proof.

Let AA be an assignment that satisfies at least 1ε1-\varepsilon fraction of the equations in EE, and consider the strategy of the provers that assigns their variables according to AA. Note that whenver each one of the equations e1,,eke_{1},\ldots,e_{k} the verifier chose is satisfied by AA, the verifier accepts. By the union bound, the probability this happens is at least 1kε1-k\varepsilon. ∎

Next, we establish the soundness of the game Gβ,rkG_{\beta,r}^{\otimes k}.

Claim 3.2.

If there is no assignment to (X,E)(X,E) satisfying at least 1ε1-\varepsilon of the equations, then the provers can win Gβ,rkG_{\beta,r}^{\otimes k} with probability at most 2Ω(ε2qrβk)2^{-\Omega(\varepsilon^{2}q^{-r}\beta k)}.

Proof.

We appeal to the parallel repetition theorem for projection games of Rao [Rao11], but we have to do so carefully. That theorem states that if Ψ\Psi is a 22-Prover-11-Round game with 𝗏𝖺𝗅(Ψ)1η{\sf val}(\Psi)\leqslant 1-\eta, then 𝗏𝖺𝗅(Ψk)2Ω(η2k){\sf val}(\Psi^{\otimes k})\leqslant 2^{-\Omega(\eta^{2}k)}. We cannot apply the theorem directly on Gβ,rG_{\beta,r} (as the square is too costly for us). Instead, we consider the game Ψ=Gβ,rqrβ\Psi=G_{\beta,r}^{\frac{q^{r}}{\beta}} and note that it has value bounded away from 11.

Write 𝗏𝖺𝗅(Ψ)=1η{\sf val}(\Psi)=1-\eta. Note that probability at least 0.10.1 there exists at least a single coordinate ii in which UiViU_{i}\neq V_{i} and all of the advice vectors v1i,,vriv_{1}^{i},\ldots,v_{r}^{i} and u1i,,uriu_{1}^{i},\ldots,u_{r}^{i} are all 0. Thus, there exists a coordinate ii and a fixing for the questions of the provers outside ii so that the answers of the players to the iith coordinate win the standard equation versus variable game with probability at least 110η1-10\eta. It follows that 110η1ε31-10\eta\leqslant 1-\frac{\varepsilon}{3}, and so ηε30\eta\geqslant\frac{\varepsilon}{30}.

We conclude from Rao’s parallel repetition theorem that

𝗏𝖺𝗅(Gβ,rk)=𝗏𝖺𝗅(Ψkqr/β)2Ω(ε2qrβk).{\sf val}(G_{\beta,r}^{\otimes k})={\sf val}(\Psi^{\otimes\frac{k}{q^{r}/\beta}})\leqslant 2^{-\Omega(\varepsilon^{2}q^{-r}\beta k)}.\qed
Viewing the advice as subspaces.

Due to the fact that each variable appears in at most O(1)O(1) equations, it can easily be seen that with probability 1O(k2/n)1-O(k^{2}/n), all variables in e1,,eke_{1},\ldots,e_{k} are distinct. In that case, note that the rr vectors of advice to the second prover, v1,,vr𝔽qVv_{1},\ldots,v_{r}\in\mathbb{F}_{q}^{V}, are uniform, and the second prover may consider their span QVQ_{V}. Note that the distribution of QVQ_{V} is that of a uniform rr dimensional subspace of 𝔽qV\mathbb{F}_{q}^{V}. As for the second prover, the vectors u1,,ur𝔽qUu_{1},\ldots,u_{r}\in\mathbb{F}_{q}^{U} are not uniformly distributed. Nevertheless, as shown by the covering property from [KS13, KMS17] (and presented below), the distribution of u1,,uru_{1},\ldots,u_{r} is close to uniform over rr-tuple of vectors from 𝔽qU\mathbb{F}_{q}^{U}. Thus, the first prover can also take their span, call it QUQ_{U}, and think of it as a random rr-dimensional subspace of 𝔽qU\mathbb{F}_{q}^{U} (which is highly correlated to QVQ_{V}).

4 The Composed PCP Construction

In this section we describe the final PCP construction, which is a composition of the outer PCP from Section 3 with the inner PCP based on the Grassmann consistency test.

4.1 The Underlying Graph

Our instance of 22-Prover-11-Round Games starts from an instance (X,𝖤𝗊)(X,{\sf Eq}) of 𝖦𝖺𝗉𝟥𝖫𝗂𝗇{\sf Gap3Lin}. Consider the game Gβ,rkG_{\beta,r}^{\otimes k} from Section 3, and let 𝒰\mathcal{U} denote the set of questions asked to the first prover. Thus 𝒰\mathcal{U} consists of all kk-tuples of equations U=(e1,,ek)𝖤𝗊kU=(e_{1},\ldots,e_{k})\in{\sf Eq}^{k} from the 𝖦𝖺𝗉𝟥𝖫𝗂𝗇{\sf Gap3Lin} instance (X,𝖤𝗊)(X,{\sf Eq}). For e𝖤𝗊e\in{\sf Eq} let ve𝔽qXv_{e}\in\mathbb{F}_{q}^{X} denote the indicator vector on the three variables appearing in ee.

It will be convenient to only keep the U=(e1,,ek)U=(e_{1},\ldots,e_{k}) that satisfy the following properties:

  • The equations e1,,eke_{1},\ldots,e_{k} are distinct and do not share variables.

  • For any iji\neq j and pair of variables xeix\in e_{i} and yejy\in e_{j}, the variables xx and yy do not appear together in any equation in the instance (X,𝖤𝗊)(X,{\sf Eq}).

The fraction of U=(e1,,ek)U=(e_{1},\ldots,e_{k}) that do not satisfy the above is O(k2/n)O(k^{2}/n) which is negligible for us, and dropping them will only reduce our completeness by o(1)o(1). This will not affect our analysis, and henceforth we will assume that all U=(e1,,ek)U=(e_{1},\ldots,e_{k}) satisfy the above properties. We now describe the 22-Prover-11-Round Games instance Ψ=(𝒜,,E,Σ1,Σ2,Φ)\Psi=(\mathcal{A},\mathcal{B},E,\Sigma_{1},\Sigma_{2},\Phi). All vertices in the underlying graph will correspond to subspaces of 𝔽qX\mathbb{F}_{q}^{X}.

4.1.1 The Vertices

For each question U=(e1,,ek)U=(e_{1},\ldots,e_{k}), let HU=span(ve1,,vek)H_{U}=\operatorname{span}(v_{e_{1}},\ldots,v_{e_{k}}), where veiv_{e_{i}} is the vector with ones at coordinates corresponding to variables appearing in eie_{i}. We can think of the veiv_{e_{i}}’s as vectors from an underlying space 𝔽qU\mathbb{F}_{q}^{U}. By the first property described above, dim(HU)=k\dim(H_{U})=k and dim(𝔽qU)=3k\dim(\mathbb{F}_{q}^{U})=3k. The vertices of Ψ\Psi are:

𝒜={LHU|U𝒰,L𝔽qU,dim(L)=2,LHU={0}},\displaystyle\mathcal{A}=\{L\oplus H_{U}\;|\;U\in\mathcal{U},L\subseteq\mathbb{F}_{q}^{U},\dim(L)=2\ell,L\cap H_{U}=\{0\}\},
={R|U𝒰, s.t. R𝔽qU,dim(R)=2(1δ)}.\displaystyle\mathcal{B}=\{R\;|\;\exists U\in\mathcal{U},\text{ s.t. }R\subseteq\mathbb{F}_{q}^{U},\dim(R)=2(1-\delta)\ell\}.

In words, the vertices on the side 𝒜\mathcal{A} are all 22\ell-dimensional subspaces of some 𝔽qU\mathbb{F}_{q}^{U} for some U𝒰U\in\mathcal{U}. For technical reasons, we require them to intersect HUH_{U} trivially (which is the case for a typical 22\ell-dimensional space) and add to them the space HUH_{U}.777This has the effect of collapsing LL and LL^{\prime} such that LHU=LHUL\oplus H_{U}=L^{\prime}\oplus H_{U} to a single vertex. The vertices on the side \mathcal{B} are all 2(1δ)2(1-\delta)\ell dimensional subspaces of 𝔽qU\mathbb{F}_{q}^{U}.

4.1.2 The Alphabets

The alphabets Σ1,Σ2\Sigma_{1},\Sigma_{2} have sizes |Σ1|=q2|\Sigma_{1}|=q^{2\ell} and |Σ2|=q2(1δ)|\Sigma_{2}|=q^{2(1-\delta)\ell}. For each vertex LHU𝒜L\oplus H_{U}\in\mathcal{A}, let ψ:HU𝔽q\psi:H_{U}\xrightarrow[]{}\mathbb{F}_{q} denote the function that satisfies the side conditions given by the equations in UU. Namely, if eiUe_{i}\in U is the equation x,hi=bi\langle x,h_{i}\rangle=b_{i} for x𝔽qUx\in\mathbb{F}_{q}^{U}, then ψ(hi)=bi\psi(h_{i})=b_{i}. We say a linear function f:LHU𝔽qf\colon L\oplus H_{U}\to\mathbb{F}_{q} satisfies the side conditions of UU if f|HUψf|_{H_{U}}\equiv\psi. In this language, for a vertex LHUL\oplus H_{U} we identify Σ1\Sigma_{1} with

{f:LHU𝔽q|f is linear function satisfying the side conditions of U}.\{f:L\oplus H_{U}\xrightarrow[]{}\mathbb{F}_{q}\;|\;f\text{ is linear function satisfying the side conditions of $U$}\}.

As LHU={0}L\cap H_{U}=\{0\} and dim(L)=2\dim(L)=2\ell, it is easy to see that the above set indeed has size q2q^{2\ell}. For each right vertex RR, we identify Σ2\Sigma_{2} with

{f:R𝔽q|f is linear}.\{f:R\xrightarrow[]{}\mathbb{F}_{q}\;|\;f\text{ is linear}\}.

4.1.3 The Edges

To define the edges, we first need the following relation on the vertices in AA. Say that (LHU)(LHU)(L\oplus H_{U})\sim(L^{\prime}\oplus H_{U^{\prime}}) if

LHUHU=LHUHU.L\oplus H_{U}\oplus H_{U^{\prime}}=L^{\prime}\oplus H_{U}\oplus H_{U^{\prime}}.

Recall that all subspaces above are in 𝔽qX\mathbb{F}_{q}^{X} hence the direct sums and equality above are well defined. The relation described is in fact an equivalence relation and thus partitions the vertices in 𝒜\mathcal{A} into disjoint equivalence classes. It is clear that the relation is reflexive and symmetric, so we need only show that it is also transitive.

Lemma 4.1.

If L1HU1HU2=L2HU1HU2L_{1}\oplus H_{U_{1}}\oplus H_{U_{2}}=L_{2}\oplus H_{U_{1}}\oplus H_{U_{2}}, and L2HU2HU3=L3HU2HU3L_{2}\oplus H_{U_{2}}\oplus H_{U_{3}}=L_{3}\oplus H_{U_{2}}\oplus H_{U_{3}}, then

L1HU1HU3=L3HU1HU3.L_{1}\oplus H_{U_{1}}\oplus H_{U_{3}}=L_{3}\oplus H_{U_{1}}\oplus H_{U_{3}}.
Proof.

We “add” HU1H_{U_{1}} to the second equation to obtain,

L2HU1HU2HU3=L3HU1HU2HU3.L_{2}\oplus H_{U_{1}}\oplus H_{U_{2}}\oplus H_{U_{3}}=L_{3}\oplus H_{U_{1}}\oplus H_{U_{2}}\oplus H_{U_{3}}.

Next, write HU2=ABH_{U_{2}}=A\oplus B, where AA is the span of all vectors vev_{e} for equations ee in U2U_{2} that are also in U1U_{1} or U3U_{3}, while BB is the span of all vectors vev_{e} for equations eU2e\in U_{2} that are in neither U1U_{1} nor U3U_{3}. It follows that AB={0}A\cap B=\{0\}. Now note that any equation eBe\in B has at most one variable that appears in an equation in U1U_{1}, and at most one variable that appears in an equation in U2U_{2}. Thus, each eBe\in B, has a “private variable”, and as the equations in BB are over disjoint sets of variables, this private variable does not appear in U1U3(U2e)U_{1}\cup U_{3}\cup(U_{2}\setminus e). It follows that

BL1L3HU1HU3={0}𝔽qX.B\cap L_{1}\oplus L_{3}\oplus H_{U_{1}}\oplus H_{U_{3}}=\{0\}\subset\mathbb{F}_{q}^{X}.

Indeed, by the above discussion any nonzero vector in B𝔽qXB\subseteq\mathbb{F}_{q}^{X} is nonzero on at least one coordinate of XX (corresponding to a private variable), and no vector in 𝔽qU1\mathbb{F}_{q}^{U_{1}} or 𝔽qU2\mathbb{F}_{q}^{U_{2}} is supported on this coordinate.

Substituting HU2=ABH_{U_{2}}=A\oplus B into the original equation yields,

L1(HU1HU3A)B=L3(HU1HU3A)B.L_{1}\oplus(H_{U_{1}}\oplus H_{U_{3}}\oplus A)\oplus B=L_{3}\oplus(H_{U_{1}}\oplus H_{U_{3}}\oplus A)\oplus B.

Since AHU1HU3A\subset H_{U_{1}}\oplus H_{U_{3}}, this equivalent to

L1HU1HU3B=L3HU1HU3B.L_{1}\oplus H_{U_{1}}\oplus H_{U_{3}}\oplus B=L_{3}\oplus H_{U_{1}}\oplus H_{U_{3}}\oplus B.

As BL1L3HU1HU3={0}B\cap L_{1}\oplus L_{3}\oplus H_{U_{1}}\oplus H_{U_{3}}=\{0\}, it follows that

L1HU1HU3=L3HU1HU3,L_{1}\oplus H_{U_{1}}\oplus H_{U_{3}}=L_{3}\oplus H_{U_{1}}\oplus H_{U_{3}},

as desired. ∎

By Lemma 4.1 the relation \sim is indeed an equivalence relation and we may partition 𝒜\mathcal{A} into equivalence classes, [LHU][L\oplus H_{U}]. We call each class a clique and partition 𝒜\mathcal{A} into cliques:

𝒜=Clique1Cliquem.\mathcal{A}=\textsf{Clique}_{1}\sqcup\cdots\sqcup\textsf{Clique}_{m}.

The actual number of cliques, mm, will not be important, but it is clear that such a number exists. The edges of our graph will be between vertices LHUL\oplus H_{U} and RR if there exists LHU[LHU]L^{\prime}\oplus H_{U^{\prime}}\in[L\oplus H_{U}] such that LRL^{\prime}\supseteq R. The edges will be weighted according to a sampling process that we describe in the next section, which also explains the constraints on Ψ\Psi. For future reference, the following lemma will be helpful in defining the constraints:

Lemma 4.2.

Suppose LHULHUL\oplus H_{U}\sim L^{\prime}\oplus H_{U^{\prime}} and that f:LHU𝔽qf:L\oplus H_{U}\xrightarrow[]{}\mathbb{F}_{q} is a linear function satisfying the side conditions. Then there is a unique linear function f:LHU𝔽qf^{\prime}:L^{\prime}\oplus H_{U^{\prime}}\xrightarrow[]{}\mathbb{F}_{q} that satisfies the side conditions such that there exists a linear function g:LHUHU𝔽qg:L\oplus H_{U}\oplus H_{U^{\prime}}\xrightarrow[]{}\mathbb{F}_{q} satisfying the side conditions (of both UU and UU^{\prime}) such that

g|LHU=f and g|LHU=f.g|_{L\oplus H_{U}}=f\quad\text{ and }\quad g|_{L^{\prime}\oplus H_{U^{\prime}}}=f^{\prime}.

In words, gg is a linear extension of both ff and ff^{\prime}.

Proof.

Note that there is only one way to extend ff to LHUHUL\oplus H_{U}\oplus H_{U^{\prime}} in a manner that satisfies the side conditions given by UU^{\prime}. Let this function be gg. We take ff^{\prime} to be g|LHUg|_{L^{\prime}\oplus H_{U^{\prime}}}. ∎

4.1.4 The Constraints

Suppose that T1T_{1} is an assignment to 𝒜\mathcal{A} that assigns, to each vertex LHUL\oplus H_{U}, a linear function T1[LHU]T_{1}[L\oplus H_{U}] satisfying the side conditions. Further suppose that T2T_{2} is an assignment that assigns to each vertex RR\in\mathcal{B} a linear function on RR. The verifier performs the following test, which also describes the constraints of Ψ\Psi:

  1. 1.

    Choose UU uniformly at random from 𝒰\mathcal{U}.

  2. 2.

    Choose LHUL\oplus H_{U} uniformly, where dim(L)=2\dim(L)=2\ell and LHU={0}L\cap H_{U}=\{0\}, and choose RLR\subseteq L of dimension 2(1δ)2(1-\delta)\ell uniformly.

  3. 3.

    Choose LHU[LHU]L^{\prime}\oplus H_{U^{\prime}}\in[L\oplus H_{U}] uniformly

  4. 4.

    As in Lemma 4.2, extend T1[LHU]T_{1}[L^{\prime}\oplus H_{U^{\prime}}] to LHUHUL^{\prime}\oplus H_{U^{\prime}}\oplus H_{U} in the unique manner that respects the side conditions and let T~1[LHU]\tilde{T}_{1}[L\oplus H_{U}] be the restriction of this extension to LHUL\oplus H_{U}.

  5. 5.

    Accept if and only if T~1[LHU]|R=T2[R]\tilde{T}_{1}[L\oplus H_{U}]|_{R}=T_{2}[R].

This finishes the description of our instance Ψ\Psi. It is clear that the running time and instance size is nO(k)n^{O(k)} and that the alphabet size is O(q2)O(q^{2\ell}).

Before arguing about the completeness and soundness, we will present some necessary tools. As is usually the case, showing completeness is relatively easy, and all of the tools presented are for the much more complex soundness analysis.

5 Tools for Soundness Analysis

In this section we will present all of the tools needed to analyze the soundness of our PCP.

5.1 The 22\ell versus 2(1δ)2\ell(1-\delta) subspace agreement test

We begin by discussing the 22\ell versus 2(1δ)2\ell(1-\delta) test and our decoding theorem for it. In our setting, we have a question U𝒰U\in\mathcal{U} for the first prover, and we consider the 22\ell versus 2(1δ)2\ell(1-\delta) test inside the space 𝔽qU\mathbb{F}_{q}^{U}. In our setting this test passes with probability at least εq2(1δ)\varepsilon\geqslant q^{-2\ell(1-\delta^{\prime})} (where δ\delta^{\prime} is, say δ=1000δ\delta^{\prime}=1000\delta) and we will want to use this fact to devise a strategy for the first prover. Below, we first state and prove a basic decoding theorem, and then deduce from it a quantitative better version that also incorporates the side conditions.

Let T1T_{1} be a table that assigns, to each L𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2)L\in\operatorname{{\sf Grass}}_{q}(\mathbb{F}_{q}^{U},2\ell), a linear function T1[L]:L𝔽qT_{1}[L]:L\xrightarrow[]{}\mathbb{F}_{q}, and let T2T_{2} be a table assigning to each R𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2(1δ))R\in\operatorname{{\sf Grass}}_{q}(\mathbb{F}_{q}^{U},2(1-\delta)\ell) a linear function T2[R]:R𝔽qT_{2}[R]\colon R\to\mathbb{F}_{q}. We recall that |U|=3k2|U|=3k\gg 2\ell. In this section, we show that if tables T1T_{1} and T2T_{2} are ε\varepsilon-consistent, namely

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2)R𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2(1δ))[T1[L]|R=T2[R]|RL]ε.\Pr_{\begin{subarray}{c}L\in\operatorname{{\sf Grass}}_{q}(\mathbb{F}_{q}^{U},2\ell)\\ R\in\operatorname{{\sf Grass}}_{q}(\mathbb{F}_{q}^{U},2(1-\delta)\ell)\end{subarray}}[T_{1}[L]|_{R}=T_{2}[R]\;|\;R\subseteq L]\geqslant\varepsilon.

for εq2(11000δ)\varepsilon\geqslant q^{-2\ell(1-1000\delta)}, then the table T1T_{1} must have non-trivial agreement with a linear function on some zoom-in and zoom-out combination of constant dimension. To prove that, we use Lemma 2.3 along with an idea from [BKS19].

Theorem 5.1.

Suppose that tables T1T_{1} and T2T_{2} are ε\varepsilon-consistent where εq2(11000δ)\varepsilon\geqslant q^{-2(1-1000\delta)\ell}. Then there exist subspaces QWQ\subset W and a linear function f:W𝔽qf:W\xrightarrow[]{}\mathbb{F}_{q} such that:

  1. 1.

    codim(Q)+dim(W)=10δ\operatorname{codim}(Q)+\dim(W)=\frac{10}{\delta}.

  2. 2.

    f|LT1[L]f|_{L}\equiv T_{1}[L] for Ω(ε)\Omega(\varepsilon^{\prime})-fraction of 22\ell-dimensional L𝖹𝗈𝗈𝗆[Q,W]L\in{{\sf Zoom}}[Q,W],

where ε=q2(11000δ2)\varepsilon^{\prime}=q^{-2(1-1000\delta^{2})\ell}.

Proof of Theorem 5.1.

Consider the bipartite graph GG whose sides are the vertices of 𝖦𝗋𝖺𝗌𝗌q(n,2){\sf Grass}_{q}(n,2\ell) and 𝖦𝗋𝖺𝗌𝗌q(n,(1δ)2){\sf Grass}_{q}(n,(1-\delta)2\ell), and its set of edges EE consists of pairs (L,R)(L,R) such that LRL\supseteq R. Consider the normalized adjacency operator 𝒯:L2(𝖦𝗋𝖺𝗌𝗌q(n,2))L2(𝖦𝗋𝖺𝗌𝗌q(n,2(1δ)))\mathcal{T}:L_{2}(\operatorname{{\sf Grass}}_{q}(n,2\ell))\xrightarrow[]{}L_{2}(\operatorname{{\sf Grass}}_{q}(n,2(1-\delta)\ell)) of GG, and let 𝒯\mathcal{T}^{*} be its adjoint operator.

Choose a linear function f:𝔽qn𝔽qf:\mathbb{F}_{q}^{n}\xrightarrow[]{}\mathbb{F}_{q} uniformly at random and define the (random) sets of vertices

SL,f={L𝖦𝗋𝖺𝗌𝗌q(n,2)|f|LT1[L]}andSR,f={R𝖦𝗋𝖺𝗌𝗌q(n,2(1δ))|f|R=T2[R]}.S_{L,f}=\{L\in{\sf Grass}_{q}(n,2\ell)\;|\;f|_{L}\equiv T_{1}[L]\}\quad\text{and}\quad S_{R,f}=\{R\in{\sf Grass}_{q}(n,2(1-\delta)\ell)\;|\;f|_{R}=T_{2}[R]\}.

Denote by E(SL,f,SR,f)E(S_{L,f},S_{R,f}) the set of edges with endpoints in SL,fS_{L,f} and SR,fS_{R,f}. We lower bound the expected size of E(SL,f,SR,f)E(S_{L,f},S_{R,f}) over the choice of ff. Note that for each edge (L,R)E(L,R)\in E such that T1[L]|RT2[R]T_{1}[L]|_{R}\equiv T_{2}[R], we have that (L,R)E(SL,f,SR,f)(L,R)\in E(S_{L,f},S_{R,f}) with probability q2q^{-2\ell}. Indeed, with probability q2q^{-2\ell} we have that T1[L]f|LT_{1}[L]\equiv f|_{L}, and in that case we automatically get that T2[R]=T1[L]|R(f|L)|R=f|RT_{2}[R]=T_{1}[L]|_{R}\equiv(f|_{L})|_{R}=f|_{R}. As the number of edges (L,R)(L,R) such that T1[L]|RT2[R]T_{1}[L]|_{R}\equiv T_{2}[R] is at least ε|E|\varepsilon|E|, we conclude that

𝔼f[|E(SL,f,SR,f)|]εq2|E|.\mathop{\mathbb{E}}_{f}\left[\left|E(S_{L,f},S_{R,f})\right|\right]\geqslant\varepsilon q^{-2\ell}|E|.

Note that we also have that

𝔼f[μ(SR,f)]=𝔼f[|SR,f||R|]=q2(1δ).\mathop{\mathbb{E}}_{f}[\mu(S_{R,f})]=\mathop{\mathbb{E}}_{f}\left[\frac{|S_{R,f}|}{|R|}\right]=q^{-2\ell(1-\delta)}.

Using Linearity of Expectation, we get that

𝔼f[|E(SL,f,SR,f)|12εq2δμ(SR,f)|E|]12εq2|E|,\mathop{\mathbb{E}}_{f}\left[\left|E(S_{L,f},S_{R,f})\right|-\frac{1}{2}\varepsilon q^{2\delta\ell}\mu(S_{R,f})|E|\right]\geqslant\frac{1}{2}\varepsilon q^{-2\ell}|E|,

thus there exists ff for which the random variable on the left hand side is at least 12εq2|E|\frac{1}{2}\varepsilon q^{-2\ell}|E|, and we fix ff so that

|E(SL,f,SR,f)|12εq2δμ(SR,f)|E|+12εq2|E|.\left|E(S_{L,f},S_{R,f})\right|\geqslant\frac{1}{2}\varepsilon q^{2\delta\ell}\mu(S_{R,f})|E|+\frac{1}{2}\varepsilon q^{-2\ell}|E|. (2)

We claim that SL,fS_{L,f} is not (r,ε)(r,\varepsilon^{\prime})-pseudo-random for r=10δr=\frac{10}{\delta} and ε=q2(11000δ2)\varepsilon^{\prime}=q^{-2\ell(1-1000\delta^{2})}. Suppose for the sake of contradiction that this is not the case, and that SL,fS_{L,f} is (r,ε)(r,\varepsilon^{\prime})-pseudo-random. Denote α=μ(SL,f)\alpha=\mu(S_{L,f}) and β=μ(SR,f)\beta=\mu(S_{R,f}). By Lemma 2.3 for any t4t\geqslant 4 that is a power of 22 we have

1|E||E(SL,SR)|qOt,r(1)βt1tεt1t+5q2rδαβqOt,r(1)βt1tεt1t.\frac{1}{|E|}|E(S_{L},S_{R})|\leqslant q^{O_{t,r}(1)}\beta^{\frac{t-1}{t}}\varepsilon^{\prime\frac{t-1}{t}}+5q^{-2r\delta\ell}\sqrt{\alpha\beta}\leqslant q^{O_{t,r}(1)}\beta^{\frac{t-1}{t}}\varepsilon^{\prime\frac{t-1}{t}}. (3)

In the last inequality, we used the fact that by (2)

β|E|=|SR,f||E||R||E(SL,f,SR,f)|12εq2|E|,\beta|E|=|S_{R,f}|\frac{|E|}{|R|}\geqslant|E(S_{L,f},S_{R,f})|\geqslant\frac{1}{2}\varepsilon q^{-2\ell}|E|,

so β12εq2q4\beta\geqslant\frac{1}{2}\varepsilon q^{-2\ell}\geqslant q^{-4\ell}, and thus the second term on the middle of (3) is negligible compared to the first term there. Combining (2) and (3) gives us that

12εq2δβqOt,r(1)βt1tεt1t.\frac{1}{2}\varepsilon q^{2\delta\ell}\beta\leqslant q^{O_{t,r}(1)}\beta^{\frac{t-1}{t}}\varepsilon^{\prime\frac{t-1}{t}}.

Simplifying, using the definition of ε\varepsilon^{\prime}, the fact that εq2(11000δ)\varepsilon\geqslant q^{-2\ell(1-1000\delta)} and the fact that βq4\beta\geqslant q^{-4\ell} we get

12q2δqOt,r(1)q4tq(2t+2000(δ2t1tδ)).\frac{1}{2}q^{2\delta\ell}\leqslant q^{O_{t,r}(1)}q^{\frac{4\ell}{t}}q^{\left(\frac{2}{t}+2000\left(\delta^{2}\frac{t-1}{t}-\delta\right)\right)\ell}.

Investigating the second two exponents of qq, we have that for t1δδ22t\geqslant\frac{1}{\delta-\delta^{2}}\geqslant 2,

(4t+2t+2000(δ2t1tδ))1994(δδ2).\left(\frac{4}{t}+\frac{2}{t}+2000\left(\delta^{2}\frac{t-1}{t}-\delta\right)\right)\ell\leqslant-1994(\delta-\delta^{2})\ell.

This implies that

12q2δqOδ(1)q1994(δ2δ)<1,\frac{1}{2}q^{2\delta\ell}\leqslant q^{O_{\delta}(1)}q^{1994(\delta^{2}-\delta)\ell}<1,

and contradiction. It follows that SL,fS_{L,f} is not (r,ε)(r,\varepsilon^{\prime})-pseudo-random, and unraveling the definition of not being pseudo-random gives the conclusion of the theorem. ∎

5.1.1 Finding a Large Fraction of Successful Zoom-Ins

Theorem 5.1 asserts the existence of a good pair of zoom-in and zoom-out (Q,W)(Q,W) on which the table T1T_{1} has good agreement with a global linear function. As discussed in the introduction, our argument requires a quantitatively version asserting that there is a good fraction of zoom-ins that work for us. Below, we state a strengthening of Theorem 5.1 achieving this; it easily follows from Theorem 5.1, and we defer the proof to Section B.

Theorem 5.2.

Suppose that tables T1T_{1} and T2T_{2} are ε\varepsilon-consistent for εq2(11000δ)\varepsilon\geqslant q^{-2\ell(1-1000\delta)}. Then there exist positive integers r1r_{1} and r2r_{2} satisfying r1+r2=r=10δr_{1}+r_{2}=r=\frac{10}{\delta}, such that for at least q52q^{-5\ell^{2}}-fraction of the r1r_{1}-dimensional subspaces QQ, there exists a subspace WQW\supseteq Q of codimension r2r_{2} and a linear function gQ,Wg_{Q,W} such that

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2)[gQ,W|L=T1[L]|QLW]q2(11000δ2).\Pr_{L\in\operatorname{{\sf Grass}}_{q}(\mathbb{F}_{q}^{U},2\ell)}\left[g_{Q,W}|_{L}=T_{1}[L]\;|\;Q\subseteq L\subseteq W\right]\geqslant q^{-2\ell(1-1000\delta^{2})}.
Proof.

The proof is deferred to Section B. ∎

5.1.2 Incorporating Side Conditions for Zoom-Ins

Next, we require a version of Theorem 5.2 which also takes the side conditions into account.

Theorem 5.3.

Let UU be a question to the first prover, let T1T_{1} the first prover’s table, including the side conditions, and suppose that

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}R𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2(1δ))[T1[LHU]|R=T2[R]|RL]=εq2(11000δ).\Pr_{\begin{subarray}{c}L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}\\ R\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2(1-\delta)\ell)\end{subarray}}[T_{1}[L\oplus H_{U}]|_{R}=T_{2}[R]\;|\;R\subseteq L]=\varepsilon\geqslant q^{-2(1-1000\delta)\ell}.

Then there are parameters r1r_{1} and r2r_{2} such that r1+r210δr_{1}+r_{2}\leqslant\frac{10}{\delta}, such that for at least q62q^{-6\ell^{2}} fraction of the r1r_{1}-dimensional subspaces Q𝔽qUQ\subseteq\mathbb{F}_{q}^{U}, there exists W𝔽qUW\subseteq\mathbb{F}_{q}^{U} of codimension r2r_{2} containing QHUQ\oplus H_{U}, and a global linear function gQ,W:W𝔽qg_{Q,W}:W\xrightarrow[]{}\mathbb{F}_{q} that respects the side conditions on HUH_{U} such that

PrL[gQ,W|LHU=T1[LHU]|QLW]q2(11000δ2)5.\Pr_{L}[g_{Q,W}|_{L\oplus H_{U}}=T_{1}[L\oplus H_{U}]\;|\;Q\subseteq L\subseteq W]\geqslant\frac{q^{-2(1-1000\delta^{2})\ell}}{5}.
Proof.

For any 2k2k-dimensional subspace AA such that HUA=𝔽qUH_{U}\oplus A=\mathbb{F}_{q}^{U}, let TAT_{A} be the table given by TA[L]=T1[LHU]|LT_{A}[L]=T_{1}[L\oplus H_{U}]|_{L} for all 22\ell-dimensional subspaces LAL\subseteq A. We can choose a 22\ell-dimensional subspace LL such that LHU={0}L\cap H_{U}=\{0\} by first uniformly choosing AA such that HUA=𝔽qUH_{U}\oplus A=\mathbb{F}_{q}^{U}, and then choosing LAL\subseteq A of dimension 22\ell uniformly. Thus, defining

p(A)=PrLA,RL[TA[L]|R=T2[R]],p^{\prime}(A)=\Pr_{L\subseteq A,R\subseteq L}[T_{A}[L]|_{R}=T_{2}[R]],

we have

𝔼A[p(A)]=PrL:dim(L)=2,LHU={0},RL[T1[LHU]|R=T2[R]]ε.\mathop{\mathbb{E}}_{A}[p^{\prime}(A)]=\Pr_{L:\dim(L)=2\ell,L\cap H_{U}=\{0\},R\subseteq L}[T_{1}[L\oplus H_{U}]|_{R}=T_{2}[R]]\geqslant\varepsilon.

In particular, for at least ε4\frac{\varepsilon}{4}-fraction of AA’s, we have p(A)ε4p^{\prime}(A)\geqslant\frac{\varepsilon}{4}. For such AA’s, by Theorem 5.2, there exist positive integers r1r_{1} and r2r_{2} such that for at least q52q^{-5\ell^{2}}-fraction of r1r_{1}-dimensional zoom-ins QQ, there exists a zoom-out WQW^{\prime}\supset Q of co-dimension r2r_{2} and a linear function gQ,Wg_{Q,W^{\prime}} such that,

PrQLW[TA[L]=gQ,W|L]q2(11000δ2)4.\Pr_{Q\subseteq L\subseteq W}[T_{A}[L]=g_{Q,W}|_{L}]\geqslant\frac{q^{-2(1-1000\delta^{2})\ell}}{4}.

Let W=WHUW=W^{\prime}\oplus H_{U} and let gQ,W:W𝔽qg_{Q,W}:W\xrightarrow[]{}\mathbb{F}_{q} be the unique extension of gQ,Wg_{Q,W^{\prime}} to WW satisfying the side conditions. We claim that

PrL:LHU={0}[gQ,W|LHU=T1[LHU]|QLW]q2(11000δ2)5.\Pr_{L:L\cap H_{U}=\{0\}}[g_{Q,W}|_{L\oplus H_{U}}=T_{1}[L\oplus H_{U}]\;|\;Q\subseteq L\subseteq W]\geqslant\frac{q^{-2(1-1000\delta^{2})\ell}}{5}.

Indeed, for each QLWQ\subseteq L^{\prime}\subseteq W^{\prime} there are an equal number of LL such that QLWQ\subseteq L\subseteq W and LHU=LHUL^{\prime}\oplus H_{U}=L\oplus H_{U}, so

PrL:LHU={0}[gQ,W|LHU=T1[LHU]|QLW]\displaystyle\Pr_{L:L\cap H_{U}=\{0\}}[g_{Q,W}|_{L\oplus H_{U}}=T_{1}[L\oplus H_{U}]\;|\;Q\subseteq L\subseteq W]
PrQLW[TA[L]=gQ,W|L]PrQLW[LHU{0}]\displaystyle\geqslant\Pr_{Q\subseteq L\subseteq W}[T_{A}[L]=g_{Q,W}|_{L}]-\Pr_{Q\subseteq L\subseteq W}[L\cap H_{U}\neq\{0\}]
q2(11000δ2)5.\displaystyle\geqslant\frac{q^{-2(1-1000\delta^{2})\ell}}{5}.

To conclude, we see that sampling AA and then QAQ\subseteq A of dimension r1r_{1}, we get that there is a zoom-out WW and a function gQ,Wg_{Q,W} satisfying the conditions in the theorem with probability at least ε4q52\frac{\varepsilon}{4}q^{-5\ell^{2}}. As the marginal distribution over QQ is qΩ(k)q^{-\Omega(k)}-close to uniform over all r1r_{1}-dimensional subspaces the conclusion follows. ∎

5.2 The Covering Property

In this section, we present the so called “covering property”, which is a feature of our PCP construction that allows us to move between the first prover’s distribution over 22\ell-dimensional subspaces of 𝔽qU\mathbb{F}_{q}^{U} and the second prover’s distribution over 22\ell-dimensional subspaces of 𝔽qV\mathbb{F}_{q}^{V}. Similar covering properties are shown in [KS13, KMS17]; however, obtaining the optimal quadratic-programming hardness result in Theorem 1.4 requires a stronger analysis that goes beyond the covering properties of [KS13, KMS17]. We are able to obtain a covering property with the following parameters:

k=q2(1+c),β=q2(1+2c/3),k=q^{2(1+c)\ell}\quad,\quad\beta=q^{-2(1+2c/3)\ell}, (4)

where c>0c>0 is a constant arbitrarily small relative to δ\delta.

5.2.1 The Basic Covering Property

To start, we state a basic form of the improved covering property that is required in our analysis and defer its proof to Appendix C. Fix a question U=(e1,,ek)U=(e_{1},\ldots,e_{k}) to the first prover and recall that HU=span(xe1,,xek)H_{U}=\operatorname{span}(x_{e_{1}},\ldots,x_{e_{k}}) where xeix_{e_{i}} is the vector that is one at coordinates corresponding to variables in eie_{i} and 0 elsewhere. The covering property we show will relate the following two distributions:

𝒟:\mathcal{D}:

  • Choose x1,,x2𝔽qUx_{1},\ldots,x_{2\ell}\in\mathbb{F}_{q}^{U} uniformly.

  • Output the list (x1,,x2)(x_{1},\ldots,x_{2\ell}).

𝒟:\mathcal{D}^{\prime}:

  • Choose VUV\subseteq U according to the Outer PCP.

  • Choose x1,,x2𝔽qVx^{\prime}_{1},\ldots,x^{\prime}_{2\ell}\in\mathbb{F}_{q}^{V} uniformly, and lift these vectors to 𝔽qU\mathbb{F}_{q}^{U} by inserting 0’s into the missing coordinates.

  • Choose w1,,w2HUw_{1},\ldots,w_{2\ell}\in H_{U} uniformly, and set xi=xi+wix_{i}=x^{\prime}_{i}+w_{i} for 1i21\leqslant i\leqslant 2\ell.

  • Output the list (x1,,x2)(x_{1},\ldots,x_{2\ell}).

With these two definitions, the covering property used in prior works asserted that the distribution 𝒟\mathcal{D} is statistically close to a variant of the distribution 𝒟\mathcal{D}^{\prime}. This closeness is not good enough for us, as we will want to consider events of rather small probability under 𝒟\mathcal{D} and still assert that their probability is roughly the same in 𝒟\mathcal{D}^{\prime}. First, in these earlier works, the distribution 𝒟\mathcal{D}^{\prime} was generated by a similar process to the above without the addition of the random vectors w1,,w2w_{1},\ldots,w_{2\ell} from HUH_{U}. As explained in the introduction however, this distribution is not good enough for the purpose of Theorem 1.4, and we must consider the distribution 𝒟\mathcal{D}^{\prime} above. Second, the notion of statistical closeness is too rough for us, and we show that in fact, almost all inputs xx are assigned the same probability under these two distributions up to factor 1+o(1)1+o(1).

More precisely, set η=q100100\eta=q^{-100\ell^{100}} throughout this subsection. Our covering property is the following statement:

Lemma 5.4.

Let η\eta be a parameter such that q100100η1/2q^{-100\ell^{100}}\leqslant\eta\leqslant 1/2. There exists a small set E(𝔽qU)2E\subseteq\left(\mathbb{F}_{q}^{U}\right)^{2\ell} such that both 𝒟(E)\mathcal{D}(E) and 𝒟(E)\mathcal{D}^{\prime}(E) are at most η40\eta^{40}, and for all (x1,,x2)E(x_{1},\ldots,x_{2\ell})\notin E we have

0.9D(x1,,x2)D(x1,,x2)1.1.0.9\leqslant\frac{D(x_{1},\ldots,x_{2\ell})}{D^{\prime}(x_{1},\ldots,x_{2\ell})}\leqslant 1.1.
Proof.

The proof is deferred to Appendix C.1. ∎

5.2.2 The Covering Property with Zoom-ins

Lemma 5.4 represents the most basic form of the covering property, and for out application we require a version of it that incorporates zoom-ins and advice. Namely, we will actually interested in the case where 𝒟\mathcal{D} and 𝒟\mathcal{D}^{\prime} are conditioned on some r1r_{1}-dimensional zoom-in QQ, for an arbitrary dimension r110δr_{1}\leqslant\frac{10}{\delta}. To make notation simpler, let us write x=(x1,,x2)x=(x_{1},\ldots,x_{2\ell}) and use spanr1(x)\operatorname{span}_{r_{1}}(x) to denote span(x1,,xr1)\operatorname{span}(x_{1},\ldots,x_{r_{1}}). Additionally, define

𝒟Q()=𝒟(x|spanr1(x)=Q)=𝒟({x|spanr1(x)=Q})𝒟({x|spanr1(x)=Q}).\mathcal{D}_{Q}(\mathcal{L})=\mathcal{D}(x\in\mathcal{L}\;|\;\operatorname{span}_{r_{1}}(x)=Q)=\frac{\mathcal{D}(\{x\in\mathcal{L}\;|\;\operatorname{span}_{r_{1}}(x)=Q\})}{\mathcal{D}(\{x\;|\;\operatorname{span}_{r_{1}}(x)=Q\})}.

From Lemma 5.4 we can conclude that for any (𝔽qU)2\mathcal{L}\subseteq(\mathbb{F}_{q}^{U})^{2\ell} that is not too small, the measure 𝒟Q()\mathcal{D}^{\prime}_{Q}(\mathcal{L}) is within at least a constant factor of 𝒟Q()\mathcal{D}_{Q}(\mathcal{L}) for nearly all QQ.

Lemma 5.5.

For any (𝔽qU)2\mathcal{L}\subseteq(\mathbb{F}_{q}^{U})^{2\ell}, we have

PrQ[𝒟Q()0.8𝒟Q()η20]12η20,\Pr_{Q}\left[\mathcal{D}^{\prime}_{Q}(\mathcal{L})\geqslant 0.8\cdot\mathcal{D}_{Q}(\mathcal{L})-\eta^{20}\right]\geqslant 1-2\eta^{20},

where QQ is the span of r1r_{1} uniformly random vectors in 𝔽qU\mathbb{F}_{q}^{U}.

Proof.

Throughout the proof all of the expectations and probabilities over QQ choose QQ as in the lemma statement. Let EE be the small set of points from Lemma 5.4. By assumption we have

𝔼Q[𝒟Q(E)]η40and𝔼Q[𝒟Q(E)]η40.\mathop{\mathbb{E}}_{Q}[\mathcal{D}_{Q}(E)]\leqslant\eta^{40}\quad\text{and}\quad\mathop{\mathbb{E}}_{Q}[\mathcal{D}^{\prime}_{Q}(E)]\leqslant\eta^{40}.

Thus, by Markov’s inequality, we have that with probability at least 12η201-2\eta^{20}, we have 𝒟Q(E),𝒟Q(E)η20\mathcal{D}_{Q}(E),\mathcal{D}^{\prime}_{Q}(E)\leqslant\eta^{20}. In this case we have,

spanr1(x)=QD(x)spanr1(x)=Q,xE¯D(x)\displaystyle\sum_{\operatorname{span}_{r_{1}}(x)=Q}D(x)\geqslant\sum_{\operatorname{span}_{r_{1}}(x)=Q,x\in\overline{E}}D(x) spanr1(x)=Q,xE¯0.9𝒟(x)\displaystyle\geqslant\sum_{\operatorname{span}_{r_{1}}(x)=Q,x\in\overline{E}}0.9\cdot\mathcal{D}^{\prime}(x)
=0.9(spanr1(x)=Q𝒟(x)spanr1(x)=Q,xE𝒟(x)),\displaystyle=0.9\cdot\left(\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}^{\prime}(x)-\sum_{\operatorname{span}_{r_{1}}(x)=Q,x\in E}\mathcal{D}^{\prime}(x)\right),

where we applied Lemma 5.4 in the second transition. Dividing both sides by spanr1(x)=Q𝒟(x)\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}^{\prime}(x) gives that

spanr1(x)=QD(x)spanr1(x)=Q𝒟(x)0.9(1𝒟Q(E))0.9(1η20)0.89.\frac{\sum_{\operatorname{span}_{r_{1}}(x)=Q}D(x)}{\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}^{\prime}(x)}\geqslant 0.9(1-\mathcal{D}^{\prime}_{Q}(E))\geqslant 0.9(1-\eta^{20})\geqslant 0.89. (5)

It follows that

𝒟Q()\displaystyle\mathcal{D}^{\prime}_{Q}(\mathcal{L}) =x,spanr1(x)=Q𝒟(x)spanr1(x)=Q𝒟(x)\displaystyle=\frac{\sum_{x\in\mathcal{L},\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}^{\prime}(x)}{\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}^{\prime}(x)}
0.89xE¯,spanr1(x)=Q𝒟(x)spanr1(x)=Q𝒟(x)\displaystyle\geqslant\frac{0.89\sum_{x\in\mathcal{L}\cap\overline{E},\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}^{\prime}(x)}{\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}(x)}
0.90.89xE¯,spanr1(x)=Q𝒟(x)spanr1(x)=Q𝒟(x)\displaystyle\geqslant\frac{0.9\cdot 0.89\cdot\sum_{x\in\mathcal{L}\cap\overline{E},\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}(x)}{\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}(x)}
0.8𝒟Q(L)spanr1(x)=Q,xE𝒟(x)spanr1(x)=Q𝒟(x)\displaystyle\geqslant 0.8\cdot\mathcal{D}_{Q}(L)-\frac{\sum_{\operatorname{span}_{r_{1}}(x)=Q,x\in E}\mathcal{D}(x)}{\sum_{\operatorname{span}_{r_{1}}(x)=Q}\mathcal{D}(x)}
=0.8𝒟Q(L)𝒟Q(E)\displaystyle=0.8\cdot\mathcal{D}_{Q}(L)-\mathcal{D}_{Q}(E)
0.8𝒟Q(L)η20,\displaystyle\geqslant 0.8\cdot\mathcal{D}_{Q}(L)-\eta^{20},

where we apply Equation (5) in the second transition and the assumption 𝒟Q(E)η20\mathcal{D}_{Q}(E)\leqslant\eta^{20} in the last transition. ∎

5.2.3 The Covering Property for the Advice

We will also need a similar, and simpler, version of the above lemma that applies to r1r_{1}-dimensional subspaces for some constant r1=O(δ1)r_{1}=O(\delta^{-1}). This is to handle the fact that the zoom-in QQ is sampled uniformly from 𝔽qV\mathbb{F}_{q}^{V} after VV is chosen according to the outer PCP, and then lifted to a subspace over 𝔽qU\mathbb{F}_{q}^{U}, instead of uniformly from 𝔽qU\mathbb{F}_{q}^{U}. Formally, let 𝒟r1\mathcal{D}^{\prime}_{r_{1}} denote the former distribution over r1r_{1}-dimensional subspaces Q𝔽qUQ\subseteq\mathbb{F}_{q}^{U} and let 𝒟r1\mathcal{D}_{r_{1}} denote the latter distribution over r1r_{1}-dimensional subspaces Q𝔽qUQ\subseteq\mathbb{F}_{q}^{U}. These are the same as the distributions 𝒟\mathcal{D} and 𝒟\mathcal{D}^{\prime} of the previous subsection except over (𝔽qU)r1\left(\mathbb{F}_{q}^{U}\right)^{r_{1}} instead of (𝔽qU)2\left(\mathbb{F}_{q}^{U}\right)^{2\ell}. We show the following.

Lemma 5.6.

Let 𝒬\mathcal{Q} be a set of r1r_{1}-dimensional subspaces in 𝔽qU\mathbb{F}_{q}^{U} satisfying 𝒟r1(𝒬)q1010\mathcal{D}_{r_{1}}(\mathcal{Q})\geqslant q^{-10\ell^{10}}. Then,

𝒟r1(𝒬)0.8𝒟r1(Q).\mathcal{D}^{\prime}_{r_{1}}(\mathcal{Q})\geqslant 0.8\cdot\mathcal{D}_{r_{1}}(Q).
Proof of Lemma 5.6.

Take EE from Lemma 5.4, and define

={(x1,,x2)(𝔽qU)2|span(x1,,xr1)𝒬}.\mathcal{L}=\{(x_{1},\ldots,x_{2\ell})\in\left(\mathbb{F}_{q}^{U}\right)^{2\ell}\;|\;\operatorname{span}(x_{1},\ldots,x_{r_{1}})\in\mathcal{Q}\}.

Then 𝒟r1(𝒬)=𝒟()\mathcal{D}_{r_{1}}(\mathcal{Q})=\mathcal{D}(\mathcal{L}), 𝒟r1(𝒬)=𝒟()\mathcal{D}^{\prime}_{r_{1}}(\mathcal{Q})=\mathcal{D}^{\prime}(\mathcal{L}) and

𝒟r1(𝒬)=𝒟()xxE𝒟(x)0.9xxE𝒟(x)0.9𝒟()𝒟(E)0.8𝒟()=0.8𝒟r1().\mathcal{D}^{\prime}_{r_{1}}(\mathcal{Q})=\mathcal{D}^{\prime}(\mathcal{L})\geqslant\sum_{\begin{subarray}{c}x\in\mathcal{L}\\ x\notin E\end{subarray}}\mathcal{D}^{\prime}(x)\geqslant 0.9\cdot\sum_{\begin{subarray}{c}x\in\mathcal{L}\\ x\notin E\end{subarray}}\mathcal{D}(x)\geqslant 0.9\cdot\mathcal{D}(\mathcal{L})-\mathcal{D}(E)\geqslant 0.8\cdot\mathcal{D}(\mathcal{L})=0.8\cdot\mathcal{D}_{r_{1}}(\mathcal{L}).

where we use Lemma 5.4 in the second transition, and the fact that 𝒟(E)η\mathcal{D}(E)\leqslant\eta and 𝒟()=𝒟r1(𝒬)q1010\mathcal{D}(\mathcal{L})=\mathcal{D}_{r_{1}}(\mathcal{Q})\geqslant q^{-10\ell^{10}} in the penultimate transition. ∎

5.3 The Number of Maximal Zoom-Outs is Bounded

In Theorem 5.2, we showed that the two provers can agree on a zoom-in with reasonable probability using their advice. The same cannot be said for zoom-outs however, and to circumvent this issue we must develop further tools. In this section, we define the notion of maximal zoom-outs and show that for a fixed zoom-in QQ, the number of maximal zoom-outs is bounded.

5.3.1 Generic Sets of Subspaces

One of our primary concerns with respect to zoom-outs is that it is possible for a prover to have many good zoom-outs to choose from (so that independent sampling doesn’t work) but not enough to allow for advice-type solution. To deal with large collections of zoom-outs we define a special property of zoom-outs that is called “genericness”. To motivate it, note that if W1,W2VW_{1},W_{2}\subseteq V are distinct subspaces of co-dimension rr, then W1W2W_{1}\cap W_{2} is a subspace whose co-dimension is between 2r2r and r+1r+1. For a typical pair of subspaces, the intersection W1W2W_{1}\cap W_{2} has dimension 2r2r, in which case we say they are generic. Genericness is useful probabilistically, since if W1,W2W_{1},W_{2} are generic, then the event that a randomly chosen 22\ell-dimensional subspace is contained in W1W_{1}, and the event it is contained in W2W_{2}, are almost independent. Below is a more general and formal definition:

Definition 5.7.

We say that a set 𝒮={W1,,WN}\mathcal{S}=\{W_{1},\ldots,W_{N}\} of codimension rr subspaces of VV is tt-generic with respect to VV if for any tt-distinct subspaces, say Wi1,,Wit𝒮W_{i_{1}},\ldots,W_{i_{t}}\in\mathcal{S}, we have codim(1jtWij)=tr\operatorname{codim}(\bigcap_{1\leqslant j\leqslant t}W_{i_{j}})=t\cdot r. When the ambient space VV is clear from context we simply say that 𝒮\mathcal{S} is tt-generic.

We remark that any set of subspaces that is tt-generic with respect to VV is also tt^{\prime}-generic with respect to VV for any t<tt^{\prime}<t. In this section, we will show a couple of results regarding generic sets of subspaces that will be used to bound the number of maximal zoom-outs in Section 8.

The result we need is a sunflower-type lemma, stating that any large set of codimension rr subspaces inside VV contains a large set of subspaces that are tt-generic with respect to VV^{\prime} for some VVV^{\prime}\subseteq V. Below is a formal statement.

Lemma 5.8.

Let t,rt,r\in\mathbb{N} be integers and let 𝒮={W1,,WN}\mathcal{S}=\{W_{1},\ldots,W_{N}\} be a set of NN subspaces of co-dimension rr inside of VV. Then there exists a subspace VVV^{\prime}\subseteq V and a set of subspaces 𝒮𝒮\mathcal{S^{\prime}}\subseteq\mathcal{S} such that:

  • |𝒮|N1(r+1)(t1)!qr\left|\mathcal{S^{\prime}}\right|\geqslant\frac{N^{\frac{1}{(r+1)\cdot(t-1)!}}}{q^{r}}.

  • Each Wi𝒮W_{i}\in\mathcal{S^{\prime}} is contained in VV^{\prime} and has co-dimension ss with respect to VV^{\prime}, where srs\leqslant r.

  • 𝒮\mathcal{S^{\prime}} is tt-generic with respect to VV^{\prime}.

In order to show Lemma 5.8, we introduce two necessary lemmas. The first, Lemma 5.9, states that for j2j\geqslant 2, any jj-generic set of subspaces contains a large (j+1)(j+1)-generic set of subspaces. The second, Lemma 5.10,states that either a set of subspaces is already 22-generic, or there are many subspaces in the set that are contained in the same hyperplane. Using this lemma, we can start from a large set of subspaces 𝒮\mathcal{S} inside of an ambient space VV and iteratively reduce to the dimension of the ambient space until we find a 22-generic set of subspaces relative to the ambient space. Indeed, either the set 𝒮\mathcal{S} is already 22-generic, or there is a hyperplane VVV^{\prime}\subseteq V such that the set of subspaces in 𝒮\mathcal{S} contained in VV^{\prime} is large. Taking this set to be the new 𝒮\mathcal{S} and VV^{\prime} to be the new ambient space, we obtain a, still, large set of subspaces, whose codimension is now one less. We may repeat this process until 𝒮\mathcal{S} is a set of hyperplanes in the ambient space, at which point it will be 22-generic.

Lemma 5.9.

Let 𝒮={W1,,𝒲N}\mathcal{S}=\{W_{1},\ldots,\mathcal{W}_{N}\} be a set of NN-subspaces of codimension rr inside of VV that is jj-generic with respect to VV, then there is a subset {W1,,WN}𝒮\{W_{1},\ldots,W_{N^{\prime}}\}\subseteq\mathcal{S} of size NN1/jqrN^{\prime}\geqslant\frac{N^{1/j}}{q^{r}} that is (j+1)(j+1)-generic with respect to VV.

Proof.

Fix any jj distinct subspaces in 𝒮\mathcal{S}, say W1,,WjW_{1},\ldots,W_{j} and let W=W1WjW=W_{1}\cap\cdots\cap W_{j}. Since 𝒮\mathcal{S} is jj-generic, codim(W)=jr\operatorname{codim}(W)=j\cdot r. We claim that there are at most qjrq^{j\cdot r} subspaces Wi𝒮{W1,,Wj}W_{i^{\prime}}\in\mathcal{S}\setminus\{W_{1},\ldots,W_{j}\} such that codim(WWi)(j+1)r1\operatorname{codim}(W\cap W_{i^{\prime}})\leqslant(j+1)r-1. Call such subspaces bad and suppose for the sake of contradiction that there are greater than qjrq^{jr} bad subspaces WiW_{i^{\prime}}. Then for each bad WiW_{i^{\prime}} we have,

dim(WiW)\displaystyle\dim(W_{i^{\prime}}\oplus W) =dim(Wi)+dim(W)dim(WiW)\displaystyle=\dim(W_{i^{\prime}})+\dim(W)-\dim(W_{i^{\prime}}\cap W)
(dim(V)r)+(dim(V)jr)(dim(V)(j+1)r1)\displaystyle\leqslant(\dim(V)-r)+(\dim(V)-jr)-(\dim(V)-(j+1)r-1)
=dim(V)1.\displaystyle=\dim(V)-1.

Therefore, for each WiW_{i^{\prime}}, the space WWiW\oplus W_{i^{\prime}} is contained in a hyperplane HH such that HWH\supseteq W. There are at most qcodim(W)1=qjr1q^{\operatorname{codim}(W)}-1=q^{j\cdot r}-1 hyperplanes HH containing WW, and by the pigenhole principle it follows that there are two bad subspaces say Wi1,Wi2W_{i^{\prime}_{1}},W_{i^{\prime}_{2}} that are both contained in the same hyperplane HH. This is a contradiction however, as by the jj-genericness of 𝒮\mathcal{S}, we must have

dim(Wi1Wi2)\displaystyle\dim(W_{i^{\prime}_{1}}\oplus W_{i^{\prime}_{2}}) =dim(Wi1)+dim(Wi2)dim(Wi1Wi2)\displaystyle=\dim(W_{i^{\prime}_{1}})+\dim(W_{i^{\prime}_{2}})-\dim(W_{i^{\prime}_{1}}\cap W_{i^{\prime}_{2}})
=2(dim(V)r)(dim(V)2r)\displaystyle=2(\dim(V)-r)-(\dim(V)-2r)
=dim(V),\displaystyle=\dim(V),

and hence Wi1W_{i^{\prime}_{1}} and Wi2W_{i^{\prime}_{2}} cannot both be contained in the hyperplane HH.

The lemma now follows from the claim we have just shown. Construct a subset 𝒮\mathcal{S^{\prime}} greedily as follows:

  1. 1.

    Initialize 𝒮\mathcal{S^{\prime}} by picking jj arbitrary subspaces from 𝒮\mathcal{S} and inserting them to 𝒮\mathcal{S^{\prime}}.

  2. 2.

    For any jj subspaces in 𝒮\mathcal{S^{\prime}}, say W1,,WjW_{1},\ldots,W_{j}, remove any W𝒮W^{\prime}\in\mathcal{S} which is bad for them.

  3. 3.

    If 𝒮\mathcal{S} is not empty, pick some W𝒮W\in\mathcal{S}, insert it to 𝒮\mathcal{S^{\prime}} and iterate.

Note that trivially, the collection 𝒮\mathcal{S^{\prime}} will be (j+1)(j+1)-generic in the end of the process. To lower bound the size of 𝒮\mathcal{S^{\prime}}, note that when |𝒮|=s|\mathcal{S^{\prime}}|=s, the number of elements from 𝒮\mathcal{S} that have been deleted is at most sjqjrs^{j}q^{jr}, and hence so long as this value is at most NN, we may do another iteration. Thus, we must have that s(Nqjr)1/j=N1/jqrs\geqslant\left(\frac{N}{q^{jr}}\right)^{1/j}=\frac{N^{1/j}}{q^{r}} when the process terminates. ∎

Lemma 5.10.

Let {W1,,WN}\{W_{1},\ldots,W_{N}\} be a set of subspaces of VV of codimension rr. Then for any integer m1m\geqslant 1, at least one of the following holds.

  • There are mm subspaces, way W1,,WmW_{1},\ldots,W_{m} such that for every pair 1ijm1\leqslant i\neq j\leqslant m, codim(WiWj)=2r\operatorname{codim}(W_{i}\cap W_{j})=2r.

  • There is a subspace VVV^{\prime}\subseteq V of co-dimension 11 that contains N=NmqrN^{\prime}=\frac{N}{mq^{r}} of these subspaces, say W1,,WNW_{1},\ldots,W_{N^{\prime}}.

Proof.

Note that for any 1ijN1\leqslant i\neq j\leqslant N, we have codim(WiWj)2r\operatorname{codim}(W_{i}\cap W_{j})\leqslant 2r. Consider the graph with vertices W1,,WNW_{1},\ldots,W_{N} with (Wi,Wj)(W_{i},W_{j}) an edge if and only if iji\neq j and codim(WiWj)2r1\operatorname{codim}(W_{i}\cap W_{j})\leqslant 2r-1. If every vertex in this graph has degree at most Nm\frac{N}{m}, then we are done as there is an independent set of size mm and these subspaces satisfy the first condition. Suppose this is not the case. Then there is a vertex, say WNW_{N}, that has Nm\frac{N}{m} neighbors, say W1,,WNmW_{1},\ldots,W_{\frac{N}{m}}. For 1iNm1\leqslant i\leqslant\frac{N}{m}, we have codim(WNWi)2r1\operatorname{codim}(W_{N}\cap W_{i})\leqslant 2r-1, so

dim(WiWN)=dim(Wi)+dim(WN)dim(WiWN)dim(V)1.\dim(W_{i}\oplus W_{N})=\dim(W_{i})+\dim(W_{N})-\dim(W_{i}\cap W_{N})\leqslant\dim(V)-1.

Thus WiWNW_{i}\oplus W_{N} is always contained in a codimension 11 subspace of VV that contains WNW_{N}. Since the number of such subspaces is qr1q^{r}-1, there must exist one subspace, say VV^{\prime}, that contains at least Nmqr\frac{N}{mq^{r}} of the subspaces in the list W1,,WNmW_{1},\ldots,W_{\frac{N}{m}} . ∎

Repeatedly applying Lemma 5.10 yields the following corollary.

Corollary 5.11.

Let {W1,,WN}\{W_{1},\ldots,W_{N}\} be a set of subspaces of VV of codimension rr with respect to VV. There exists a subspace VVV^{\prime}\subset V, an integer 1sr1\leqslant s\leqslant r, and a subset of mN1r+1qrm\geqslant\frac{N^{\frac{1}{r+1}}}{q^{r}}, say {W1,,Wm}\{W_{1},\ldots,W_{m}\}, all contained in VV^{\prime} such that,

  • Each WiW_{i}, 1im1\leqslant i\leqslant m, has codimension ss with respect to VV^{\prime}.

  • Each WiWjW_{i}\cap W_{j}, 1ijm1\leqslant i\neq j\leqslant m, has codimension 2s2s.

Proof.

To start set V=VV^{\prime}=V. If the WiW_{i}’s have codimension 11 in VV^{\prime} then the result holds.

Otherwise, if the conclusion does not hold, then apply Lemma 5.10 with m=N1r+1qrm=\frac{N^{\frac{1}{r+1}}}{q^{r}}. Either the first condition of Lemma 5.10 holds and we are done, or we can find a new subspace, V′′V^{\prime\prime}, of codimension 11 inside the current VV^{\prime} containing at least Nmqr\frac{N}{mq^{r}} of the WiW_{i}’s. Set V=V′′V^{\prime}=V^{\prime\prime} and repeat. Note that the codimension of the WiW_{i}’s with respect to VV^{\prime} drops by 11 after every iteration, so we will repeat at most rr times before reaching the desired conclusion. This yields a list of WiW_{i}’s that satisfy the conditions of size at least

N(mqr)r=N1r+1m.\frac{N}{(mq^{r})^{r}}=N^{\frac{1}{r+1}}\geqslant m.\qed

With Corollary 5.11 and Lemma 5.9, we can prove Lemma 5.8

Proof of Lemma 5.8.

By Corollary 5.11, there is a set 𝒮\mathcal{S}^{\prime} of size |𝒮|N1r+1qr|\mathcal{S}^{\prime}|\geqslant\frac{N^{\frac{1}{r+1}}}{q^{r}} and VVV^{\prime}\subseteq V such that 𝒮\mathcal{S^{\prime}} is 22-generic with respect to VV^{\prime}, and each WiW_{i} has (the same) codimension srs\leqslant r with respect to VV^{\prime}. Applying Lemma 5.9 t2t-2 times, there is a set of tt-generic subspaces relative to VV^{\prime}, 𝒮′′𝒮\mathcal{S}^{\prime\prime}\subseteq\mathcal{S^{\prime}}, of size

|𝒮′′|((((|𝒮|qr)121qr)13)1qr)1t1|𝒮|1(t1)!qrN1(r+1)(t1)!qr.\left|\mathcal{S}^{\prime\prime}\right|\geqslant\left(\left(\left(\left(\frac{|\mathcal{S}^{\prime}|}{q^{r}}\right)^{\frac{1}{2}}\cdot\frac{1}{q^{r}}\right)^{\frac{1}{3}}\cdots\right)\cdot\frac{1}{q^{r}}\right)^{\frac{1}{t-1}}\geqslant\frac{\left|\mathcal{S}^{\prime}\right|^{\frac{1}{(t-1)!}}}{q^{r}}\ \geqslant\frac{N^{\frac{1}{(r+1)\cdot(t-1)!}}}{q^{r}}.\qed

In addition to Lemma 5.8, we state another useful feature of generic sets of subspaces, formalized in Lemma 5.12 below. The lemma asserts that if a collection {W1,,WN}\{W_{1},\ldots,W_{N}\} is generic, and one zooms-outs from the ambient space VV into a hyperplane HH, then one gets an induced collection {W1H,,WNH}\{W_{1}\cap H,\ldots,W_{N}\cap H\} which is almost as generic.

Lemma 5.12.

Let 𝒮={W1,,WN}\mathcal{S}=\{W_{1},\ldots,W_{N}\} be a set of subspaces of codimension rr that is tt-generic with respect to some space VV for an even integer tt, and let HH be a hyperplane in VV. Then the set of subspaces 𝒮={W1H,,WNH}\mathcal{S}^{\prime}=\{W_{1}\cap H,\ldots,W_{N}\cap H\} can be made a t2\frac{t}{2}-generic set of subspaces with respect to HH with codimension rr inside of HH by removing at most t2\frac{t}{2} subspaces WiHW_{i}\cap H from it.

Proof.

Suppose that 𝒮\mathcal{S}^{\prime} is not t2\frac{t}{2}-generic with respect to HH with codimension rr inside of HH, as otherwise we are done. In this case, there must exist t2\frac{t}{2} distinct subspaces, say W1H,,Wt2H𝒮W_{1}\cap H,\ldots,W_{\frac{t}{2}}\cap H\in\mathcal{S}^{\prime} such that

codim(W1Wt2H)<t2r+1,\operatorname{codim}\left(W_{1}\cap\cdots\cap W_{\frac{t}{2}}\cap H\right)<\frac{t}{2}\cdot r+1,

where the codimension is with respect to VV. However, since 𝒮\mathcal{S} is tt-generic (and thus t2\frac{t}{2}-generic as well) with respect to VV, this implies that

W1Wt2H.W_{1}\cap\cdots\cap W_{\frac{t}{2}}\subseteq H.

Now delete W1H,,Wt2HW_{1}\cap H,\ldots,W_{\frac{t}{2}}\cap H from 𝒮\mathcal{S}^{\prime}. We claim that the resulting set is t2\frac{t}{2}-generic with respect to HH. Suppose for the sake of contradiction that it is not. Then there must be another t2\frac{t}{2} distinct subspaces, say Wt2+1H,,WtH𝒮W_{\frac{t}{2}+1}\cap H,\ldots,W_{t}\cap H\in\mathcal{S}^{\prime} such that

Wt2+1WtH.W_{\frac{t}{2}+1}\cap\cdots\cap W_{t}\subseteq H.

This would imply

(W1Wt2)(Wt2+1Wt)H.\left(W_{1}\cap\cdots\cap W_{\frac{t}{2}}\right)\oplus\left(W_{\frac{t}{2}+1}\cap\cdots\cap W_{t}\right)\subseteq H.

This is a contradiction however, as 𝒮\mathcal{S} is tt-generic with respect to VV, so codim(W1Wt)=tr\operatorname{codim}(W_{1}\cap\cdots\cap W_{t})=tr, and

dim(W1Wt2Wt2+1Wt)\displaystyle\dim\left(W_{1}\cap\cdots\cap W_{\frac{t}{2}}\oplus W_{\frac{t}{2}+1}\cap\cdots\cap W_{t}\right)
=dim(W1Wt2)+dim(Wt2Wt2+1Wt)dim(W1W2Wt)\displaystyle\qquad\qquad=\dim\left(W_{1}\cap\cdots\cap W_{\frac{t}{2}}\right)+\dim\left(W_{\frac{t}{2}}\oplus W_{\frac{t}{2}+1}\cap\cdots\cap W_{t}\right)-\dim\left(W_{1}\oplus W_{2}\cap\cdots\cap W_{t}\right)
=2(dim(V)t2r)dim(V)+tr=dim(V)>dim(H),\displaystyle\qquad\qquad=2\left(\dim(V)-\frac{t}{2}r\right)-\dim(V)+tr=\dim(V)>\dim(H),

and contradiction. ∎

Lemma 5.13.

Let 𝒲\mathcal{W} be a set of subspaces that is 2K2^{K} generic with respect to VV and let BB be a subspace of codimension jj. Then, the set of subspaces,

𝒲B={WiB|Wi𝒲},\mathcal{W}_{B}=\{W_{i}\cap B\;|\;W_{i}\in\mathcal{W}\},

can be made 2Kj2^{K-j} generic with respect to BB by removing at most j2K1j2^{K-1} subspaces.

Proof.

There is a sequence of subspaces V=B0B1Bj=BV=B_{0}\supseteq B_{1}\supseteq\cdots\supseteq B_{j}=B, such that Bi+1B_{i+1} is a hyperplane inside of BiB_{i}. Do the following,

  1. 1.

    Initialize 𝒲0=𝒲\mathcal{W}_{0}=\mathcal{W} and set i=1i=1.

  2. 2.

    Set 𝒲i={WkBi|Wk𝒲i1}\mathcal{W}_{i}=\{W_{k}\cap B_{i}\;|\;W_{k}\in\mathcal{W}_{i-1}\}, and then remove the minimal number of subspaces to turn 𝒲i\mathcal{W}_{i} into a 2Ki2^{K-i}-generic collection with respect to BiB_{i}.

  3. 3.

    Stop if i=ji=j, otherwise, increase ii by 11 and return to step 22.

It is clear that the output is a set of subspaces 𝒲j𝒲B\mathcal{W}_{j}\subseteq\mathcal{W}_{B} that is 2Kj2^{K-j}-generic with respect to BB. Furthermore, during each iteration, at most 2Ki12K12^{K-i-1}\leqslant 2^{K-1} subspaces are removed by Lemma 5.12, and the result follows. ∎

5.3.2 The Sampling Lemma

As explained earlier, the notion of genericness is useful probabilistically, and in this section we state and prove a sampling lemma about generic collections which is necessary for our analysis. Fix an arbitrary zoom-in QVQ\subseteq V of dimension aa, and let 𝒮={W1,,Wm}\mathcal{S}=\{W_{1},\ldots,W_{m}\} be a 22-generic collection of subspaces of VV of codimension rr all containing QQ. Also let 𝒜\mathcal{A} be a set of jj-dimensional subspaces containing QQ. For the remainder of this subsection, use 𝖹𝗈𝗈𝗆[Q,V]{\sf Zoom}[Q,V] to denote the set of jj-dimensional subspaces in VV containing QQ. Consider the following two probability measures over 𝖹𝗈𝗈𝗆[Q,V]{\sf Zoom}[Q,V]:

  1. 1.

    The distribution μ\mu which is uniform over 𝖹𝗈𝗈𝗆[Q,V]{\sf Zoom}[Q,V].

  2. 2.

    The distribution ν\nu, wherein a subspace is sampled by first picking i{1,,m}i\in\{1,\ldots,m\} uniformly and then sampling a subspace from 𝖹𝗈𝗈𝗆[Q,Wi]{\sf Zoom}[Q,W_{i}] uniformly.

The main content of this section is the following lemma, asserting that the measures μ\mu and ν\nu are close in statistical distance provided that mm is large. More precisely:

Lemma 5.14.

For any 𝖹𝗈𝗈𝗆[Q,V]\mathcal{L}\subset{\sf Zoom}[Q,V] we have

|ν()μ()|3qr2(ja)m.\left|\nu(\mathcal{L})-\mu(\mathcal{L})\right|\leqslant\frac{3q^{\frac{r}{2}(j-a)}}{\sqrt{m}}.

We now set up some notations for the proof of Lemma 5.14. For L𝖹𝗈𝗈𝗆[Q,V]L\in{\sf Zoom}[Q,V] let

N(L)=|{Wi𝒮|LWi}|,N(L)=|\{W_{i}\in\mathcal{S}\;|\;L\subseteq W_{i}\}|,

and for an arbitrary pair of distinct Wi,Wi𝒮W_{i},W_{i^{\prime}}\in\mathcal{S} define the following quantities:

D=|{L𝖹𝗈𝗈𝗆[Q,V]|LWi}=[narja]q,p1=PrL𝖹𝗈𝗈𝗆[Q,V][LWi]\displaystyle D=|\{L\in{\sf Zoom}[Q,V]\;|\;L\subseteq W_{i}\}=\begin{bmatrix}{n-a-r}\\ {j-a}\end{bmatrix}_{q},\qquad p_{1}=\Pr_{L\in{\sf Zoom}[Q,V]}[L\subseteq W_{i}]
p2=PrL𝖹𝗈𝗈𝗆[Q,V][LWiWi]=[na2rja]q[naja]q.\displaystyle\qquad\qquad\qquad p_{2}=\Pr_{L\in{\sf Zoom}[Q,V]}[L\subseteq W_{i}\cap W_{i^{\prime}}]=\frac{\begin{bmatrix}{n-a-2r}\\ {j-a}\end{bmatrix}_{q}}{\begin{bmatrix}{n-a}\\ {j-a}\end{bmatrix}_{q}}. (6)

We note that all of these quantities are well defined as they do not depend on the identity of WiW_{i} and WiW_{i^{\prime}}. The first two equations are clear and the third uses the fact that 𝒮\mathcal{S} is 22-generic. Thus,

μ(L)=1|𝖹𝗈𝗈𝗆[Q,V]|=p1D,ν(L)=N(L)m1D=N(L)mD.\mu(L)=\frac{1}{|{\sf Zoom}[Q,V]|}=\frac{p_{1}}{D},\qquad\qquad\qquad\nu(L)=\frac{N(L)}{m}\cdot\frac{1}{D}=\frac{N(L)}{mD}. (7)

For the first equation, we are using the fact that |𝖹𝗈𝗈𝗆[Q,V]|p1=D|{\sf Zoom}[Q,V]|\cdot p_{1}=D, while the second equation is evident. In the following claim we analyze the expectation and variance of N(L)N(L) when LL is chosen uniformly form 𝖹𝗈𝗈𝗆[Q,V]{\sf Zoom}[Q,V]:

Claim 5.15.

𝔼L𝖹𝗈𝗈𝗆[Q,V][N(L)]=p1m\mathop{\mathbb{E}}_{L\in{\sf Zoom}[Q,V]}[N(L)]=p_{1}m and var(N(L))p1m\operatorname{var}(N(L))\leqslant p_{1}m. where the variance is over uniform L𝖹𝗈𝗈𝗆[Q,V]L\in{\sf Zoom}[Q,V].

Proof.

By linearity of expectation

𝔼L𝖹𝗈𝗈𝗆[Q,V][N(L)]=i=1mPrL𝖹𝗈𝗈𝗆[Q,V][LWi]=p1m,\mathop{\mathbb{E}}_{L\in{\sf Zoom}[Q,V]}[N(L)]=\sum_{i=1}^{m}\Pr_{L\in{\sf Zoom}[Q,V]}[L\subseteq W_{i}]=p_{1}m,

and we move on to the variance analysis. To bound 𝔼L𝖹𝗈𝗈𝗆[Q,V][N(L)2]\mathop{\mathbb{E}}_{L\in{\sf Zoom}[Q,V]}[N(L)^{2}], write

𝔼L𝖹𝗈𝗈𝗆[Q,V][N(L)2]\displaystyle\mathop{\mathbb{E}}_{L\in{\sf Zoom}[Q,V]}[N(L)^{2}] =𝔼L𝖹𝗈𝗈𝗆[Q,V][(i=1m𝟙LWi)2]\displaystyle=\mathop{\mathbb{E}}_{L\in{\sf Zoom}[Q,V]}\left[\left(\sum_{i=1}^{m}\mathbbm{1}_{L\subseteq W_{i}}\right)^{2}\right]
mPrL𝖹𝗈𝗈𝗆[Q,V][LWi]+m2PrL𝖹𝗈𝗈𝗆[Q,V][LWiWi]\displaystyle\leqslant m\cdot\Pr_{L\in{\sf Zoom}[Q,V]}[L\subseteq W_{i}]+m^{2}\cdot\Pr_{L\in{\sf Zoom}[Q,V]}[L\subseteq W_{i}\cap W_{i^{\prime}}]
=p1m+p2m2.\displaystyle=p_{1}m+p_{2}m^{2}.

It follows that,

var(N(L))p1m+p2m2p12m2.\operatorname{var}(N(L))\leqslant p_{1}m+p_{2}m^{2}-p_{1}^{2}m^{2}.

Finally note that p2p_{2} and p12p_{1}^{2} are nearly the same value, and it can be checked using Equation (6) that p2p12p_{2}\leqslant p_{1}^{2} and so var(N(U))p1m\operatorname{var}(N(U))\leqslant p_{1}m. ∎

Combining Chebyshev’s inequality with Claim 5.15, we conclude the following lemma which will be useful for us later on.

Lemma 5.16.

For any c>0c>0 it holds that

PrL𝖹𝗈𝗈𝗆[Q,V][|N(L)p1m|cp1m]1c2p1mqr(ja)c2m.\Pr_{L\in{\sf Zoom}[Q,V]}\Biggl{[}\left|N(L)-p_{1}m\right|\geqslant c\cdot p_{1}m\Biggr{]}\leqslant\frac{1}{c^{2}p_{1}m}\leqslant\frac{q^{r(j-a)}}{c^{2}m}.
Proof.

This is an immediate result of Chebyshev’s inequality with the bounds from Claim 5.15. ∎

Lastly, we use Claim 5.15 to prove Lemma 5.14.

Proof of Lemma 5.14.

We have,

|μ()ν()|=1mD|LN(L)p1m|1mDL|N(L)p1m||𝖹𝗈𝗈𝗆[Q,V]|mD𝔼L[|N(L)p1m|],|\mu(\mathcal{L})-\nu(\mathcal{L})|=\frac{1}{mD}|\sum\limits_{L\in\mathcal{L}}N(L)-p_{1}m|\leqslant\frac{1}{mD}\sum\limits_{L\in\mathcal{L}}|N(L)-p_{1}m|\leqslant\frac{|{\sf Zoom}[Q,V]|}{mD}\mathop{\mathbb{E}}_{L}[|N(L)-p_{1}m|],

and by Cauchy-Schwartz we get that

|μ()ν()||𝖹𝗈𝗈𝗆[Q,V]|mDvar(N(L)).|\mu(\mathcal{L})-\nu(\mathcal{L})|\leqslant\frac{|{\sf Zoom}[Q,V]|}{mD}\sqrt{\operatorname{var}(N(L))}.

Plugging in Claim 5.15 and using |𝖹𝗈𝗈𝗆[Q,V]|=D/p1|{\sf Zoom}[Q,V]|=D/p_{1} we get that

|μ()ν()|1p1m3qr2(ja)m.|\mu(\mathcal{L})-\nu(\mathcal{L})|\leqslant\frac{1}{\sqrt{p_{1}m}}\leqslant\frac{3q^{\frac{r}{2}(j-a)}}{\sqrt{m}}.\qed

5.3.3 Bound on Maximal Zoom-outs

For this subsection, we work in the second prover’s space, 𝔽qV\mathbb{F}_{q}^{V}, and make the assumption that |V||V|\gg\ell, say |V|2100q|V|\geqslant 2^{100}q^{\ell} to be concrete. We first establish several results in the simplified setting where there is no zoom-in. After that we show how to deduce an analogous result with a zoom-in. Throughout this section, we fix TT to be a table that assigns to each L𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell) a linear function on LL.

Definition 5.17.

Given a table TT on 𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2){\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell) and a subspace Q𝔽qVQ\subseteq\mathbb{F}_{q}^{V}, we call a zoom-out, function pair, (W,gW)(W,g_{W}), where W𝔽qVW\subseteq\mathbb{F}_{q}^{V} and f:W𝔽qf:W\xrightarrow[]{}\mathbb{F}_{q}, (C,s)(C,s)-maximal with respect to TT on QQ if

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[gW|LT[L]|QLW]C,\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[g_{W}|_{L}\equiv T[L]\;|\;Q\subseteq L\subseteq W]\geqslant C,

and there does not exist another zoom-out function pair, (W,gW)(W^{\prime},g_{W^{\prime}}) such that 𝔽qVWW\mathbb{F}_{q}^{V}\supseteq W^{\prime}\supsetneq W, gW:𝔽W𝔽qg_{W^{\prime}}:\mathbb{F}_{W}^{\prime}\xrightarrow[]{}\mathbb{F}_{q}, gW|Wg|Wg_{W^{\prime}}|_{W}\equiv g|_{W} and

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[gW|LT[L]|QLW]sC.\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[g_{W^{\prime}}|_{L}\equiv T[L]\;|\;Q\subseteq L\subseteq W^{\prime}]\geqslant sC.

In the case that Q={0}Q=\{0\}, we say that (W,gW)(W,g_{W}) is (C,s)(C,s)-maximal with respect to TT.

In the above statement, CC should be thought of as small and ss should be thought of as an absolute constant. With this in mind, a zoom-out WW and a linear function on it gWg_{W} is called maximal if there is no zoom-out WW^{\prime} that strictly contains WW, and an extension of gWg_{W} to gWg_{W^{\prime}}, that has the same agreement with TT as gWg_{W} (up to constant factors). As an immediate consequence of the definition of (C,s)(C,s)-maximal, we have the following lemma, which roughly states that every zoom-out for which there that has a linear function with good agreement inside the zoom-out is contained in a maximal zoom-out (with only slightly worse agreement).

Lemma 5.18.

Let TT be a table on 𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2){\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell), Q𝔽qVQ\subseteq\mathbb{F}_{q}^{V}, and W𝔽qVW\subseteq\mathbb{F}_{q}^{V} be a subspace of codimension rr containing QQ. Suppose that there exists a linear function gW:W𝔽qg_{W}:W\xrightarrow[]{}\mathbb{F}_{q} such that

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[gW|LT[L]|QLW]C.\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[g_{W}|_{L}\equiv T[L]\;|\;Q\subseteq L\subseteq W]\geqslant C.

Then there exists a subspace WWW^{\prime}\supseteq W and a linear function gW:W𝔽qg_{W^{\prime}}:W^{\prime}\xrightarrow[]{}\mathbb{F}_{q} such that gW|WgWg_{W^{\prime}}|_{W}\equiv g_{W} and (gW,W)(g_{W^{\prime}},W^{\prime}) is (Csr,s)(Cs^{-r},s)-maximal and a linear function gW:W𝔽qg_{W^{\prime}}:W^{\prime}\xrightarrow[]{}\mathbb{F}_{q} such that

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[gW|LT[L]|QLW]C.\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[g_{W^{\prime}}|_{L}\equiv T[L]\;|\;Q\subseteq L\subseteq W]\geqslant C.
Proof.

This is an immediate consequence of Definition 5.17. If (W,gW)(W,g_{W}) is (C,s)(C,s)-maximal then we are done. Otherwise, there must exist W1,gW1W_{1},g_{W_{1}} such that W1WW_{1}\supsetneq W, gW1|WgWg_{W_{1}}|_{W}\equiv g_{W} and

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[gW1|LT[L]|QLW1]sC.\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[g_{W_{1}}|_{L}\equiv T[L]\;|\;Q\subseteq L\subseteq W_{1}]\geqslant sC.

We can repeat this argument at most rr times before obtaining some (gW,W)(g_{W^{\prime}},W^{\prime}) that is (Csr,s)(Cs^{-r},s) maximal and satisfies WWW^{\prime}\supseteq W and gW|WgWg_{W^{\prime}}|_{W}\equiv g_{W}. ∎

The following result, which is key to our analysis gives an upper bound on the number of maximal zoom-outs. This lemma is in fact equivalent to Lemma 1.7, and its proof is deferred to Section 8. In order to present the lemma cleanly, we set the following parameters for the remainder of the section, which can all be considered constant

ξ=δ5,δ2=ξ100,t=(22+10/δ2)!.\xi=\delta^{5},\quad\delta_{2}=\frac{\xi}{100},\quad t=\left(2^{2+10/\delta_{2}}\right)!\;.
Lemma 5.19.

Let TT be a table on 𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2){\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell) with |V|2100q|V|\geqslant 2^{100}q^{\ell} and set r10δr\leqslant\frac{10}{\delta}, Cq2(1δ5)C\geqslant q^{-2(1-\delta^{5})\ell}, and

Nq100(t1)!r2ξ1.N\geqslant q^{100\left(t-1\right)!r^{2}\ell\xi^{-1}}.

Suppose that (W1,f1),,(WN,fN)(W_{1},f_{1}),\ldots,(W_{N},f_{N}) are zoom-out, function pairs such that the WiW_{i}’s are all distinct and for each 1im1\leqslant i\leqslant m, codim(Wi)=r\operatorname{codim}(W_{i})=r, and

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[fi|LT[L]|LWi]C.\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[f_{i}|_{L}\equiv T[L]\;|\;L\subseteq W_{i}]\geqslant C.

Then there is a a subspace VV^{\prime}, a linear function h:V𝔽qh^{\prime}:V^{\prime}\xrightarrow[]{}\mathbb{F}_{q}, and a set of subspaces 𝒲{W1,,WN}\mathcal{W}^{\prime}\subseteq\{W_{1},\ldots,W_{N}\} of size Nq50rξ1N\geqslant q^{50r\ell\xi^{-1}}

  • Each Wi𝒲W_{i}\in\mathcal{W}^{\prime} is strictly contained in VV^{\prime} and has codimension r<rr^{\prime}<r with respect to VV^{\prime}.

  • 𝒲\mathcal{W}^{\prime} is 22-generic with respect to VV^{\prime}.

  • For any Wi𝒲W_{i}\in\mathcal{W}^{\prime}, h|Wifih^{\prime}|_{W_{i}}\equiv f_{i}.

Proof.

The proof is deferred to Section 8. ∎

To bound the number of maximal zoom-outs, we will also need the following list decoding property.

Lemma 5.20.

Let TT be a table on 𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2){\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell), let QQ be an r1r_{1}-dimensional subspace, and let WQW\supseteq Q be a subspace of codimension r2r_{2}. Suppose that 22\ell is sufficiently large and dim(W)20\dim(W)\geqslant 20\ell. Let f1,,fmf_{1},\ldots,f_{m} be a list of distinct linear functions such that fi|LT[L]f_{i}|_{L}\equiv T[L] for at least β\beta-fraction of the 22\ell-dimensional subspaces LL such that QLWQ\subseteq L\subseteq W, for β2q2+r1+c\beta\geqslant 2q^{-2\ell+r_{1}}+c, and c>0c>0. Then,

m4c24β2.m\leqslant\frac{4}{c^{2}}\leqslant\frac{4}{\beta^{2}}.
Proof.

The proof is deferred to Appendix D

Combining Lemma 5.19 with Lemma 5.20, yields a bound on the number of (C,s)(C,s) maximal zoom-out function pairs with respect to a table TT on QQ.

Theorem 5.21.

For any table TT on 𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2){\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell) such that |V|q|V|\geqslant q^{\ell}, any subspace Q𝔽qVQ\subseteq\mathbb{F}_{q}^{V} of dimension r110δr_{1}\leqslant\frac{10}{\delta} and any Cq2(1δ3)C\geqslant q^{-2(1-\delta^{3})\ell}, the number of (C,15)(C,\frac{1}{5})-maximal zoom-out, function pairs with respect to TT on QQ is at most 40δC2q100(t1)!r2ξ1\frac{40}{\delta}\cdot C^{-2}\cdot q^{100\left(t-1\right)!r^{2}\ell\xi^{-1}}.

Proof.

Suppose for the sake of contradiction that (W1,f1),,(WM,fM)(W_{1},f_{1}),\ldots,(W_{M},f_{M}) are M>40δC2q100(t1)!r2ξ1M>\frac{40}{\delta}C^{-2}q^{100(t-1)!r^{2}\ell\xi^{-1}} distinct pairs that are (C,15)(C,\frac{1}{5}) maximal with respect to TT on QQ. By Lemma 5.20, for each WiW_{i}, there are at most 4C24C^{-2} functions f:W𝔽qf:W\xrightarrow[]{}\mathbb{F}_{q} such that f|LT[L]f|_{L}\equiv T[L] for at least CC-fraction of the L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W]. Thus, there are C2M/4C^{2}M/4 distinct WiW_{i}’s appearing in the pairs, and there is a codimension r210δr_{2}\leqslant\frac{10}{\delta} such there are M40δC2=Nq100(t1)!r2ξ1\frac{M}{\frac{40}{\delta}C^{-2}}=N\geqslant q^{100\left(t-1\right)!r^{2}\ell\xi^{-1}} pairs, say, (W1,f1),,(WN,fN)(W_{1},f_{1}),\ldots,(W_{N},f_{N}) zoom-out function pairs that are (C,15)(C,\frac{1}{5}) maximal with respect to TT on QQ such that the WiW_{i}’s are all distinct and of codimension r2r_{2} in 𝔽qV\mathbb{F}_{q}^{V}.

Write 𝔽qV=QA\mathbb{F}_{q}^{V}=Q\oplus A. For each L𝔽qVL\subseteq\mathbb{F}_{q}^{V} of dimension 22\ell containing QQ, there is a unique LAL^{\prime}\subseteq A such that L=QLL=Q\oplus L^{\prime}. Define the table TT^{\prime} that assigns linear functions to each L𝖦𝗋𝖺𝗌𝗌q(A,2dim(Q))L^{\prime}\in{\sf Grass}_{q}(A,2\ell-\dim(Q)) by

T[L]T[LQ]|LT^{\prime}[L^{\prime}]\equiv T[L^{\prime}\oplus Q]|_{L^{\prime}} (8)

For each 1iN1\leqslant i\leqslant N, let WiAW^{\prime}_{i}\subseteq A be the unique subspace such that Wi=WiQW_{i}=W^{\prime}_{i}\oplus Q. We have that fi|L=T[L]f_{i}|_{L^{\prime}}=T^{\prime}[L^{\prime}] for at least CC-fraction of L𝖦𝗋𝖺𝗌𝗌q(Wi,2dim(Q))L^{\prime}\in{\sf Grass}_{q}(W^{\prime}_{i},2\ell-\dim(Q)).

By Lemma 5.19 there exists a subspace VV^{\prime}, a linear function h:V𝔽qh^{\prime}:V^{\prime}\xrightarrow[]{}\mathbb{F}_{q}, and a subcollection 𝒲{W1,,Wm}\mathcal{W}^{\prime}\subseteq\{W^{\prime}_{1},\ldots,W^{\prime}_{m}\} of size at least mq50rξ1m\geqslant q^{50r\ell\xi^{-1}} such that

  • Each Wi𝒲W^{\prime}_{i}\in\mathcal{W}^{\prime} is has codimension 1rr21\leqslant r^{\prime}\leqslant r_{2} with respect to VV^{\prime}.

  • 𝒲\mathcal{W}^{\prime} is 22-generic with respect to VV^{\prime}.

  • For any Wi𝒲W^{\prime}_{i}\in\mathcal{W}, h|Wi=fih^{\prime}|_{W^{\prime}_{i}}=f_{i}.

Let 𝒲={Wi|Wi𝒲}\mathcal{W}=\{W_{i}\;|\;W^{\prime}_{i}\in\mathcal{W}^{\prime}\} and extend hh^{\prime} to the function hh^{\star} on V=VQV^{\star}=V^{\prime}\oplus Q so that, h|Wifih^{\star}|_{W_{i}}\equiv f_{i} for at least qr1q^{-r_{1}} of the WiW_{i} in 𝒲\mathcal{W}. It follows that there is a set 𝒱={Wi𝒲|h|Wifi}\mathcal{V}=\{W_{i}\in\mathcal{W}\;|\;h^{\star}|_{W_{i}}\equiv f_{i}\} of size |𝒱|mqr2mq10δ|\mathcal{V}|\geqslant mq^{-r_{2}}\geqslant mq^{-\frac{10}{\delta}}.

Furthermore, because 𝒲\mathcal{W}^{\prime} is 22-generic inside of VV^{\prime}, 𝒲\mathcal{W} is 22-generic inside of VV^{\star}. We will now finish the proof by applying Lemma 5.14 on 𝒱\mathcal{V}. Specifically, let ν\nu denote the measure over 𝖹𝗈𝗈𝗆[Q,W]{\sf Zoom}[Q,W^{\star}] generated by choosing Wi𝒱W_{i}\in\mathcal{V} and then L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W^{\star}] conditioned on LWiL\subseteq W_{i}, let μ\mu denote the uniform measure over 𝖹𝗈𝗈𝗆[Q,W]{\sf Zoom}[Q,W^{\star}], and let

={L𝖹𝗈𝗈𝗆[Q,V]|h|LT[L]}.\mathcal{L}=\{L\subseteq{\sf Zoom}[Q,V^{\star}]\;|\;h^{\star}|_{L}\equiv T[L]\}.

Since h|Wifih^{\star}|_{W_{i}}\equiv f_{i} for every Wi𝒱W_{i}\in\mathcal{V}, we have

ν()𝔼Wi𝒱[PrL𝖹𝗈𝗈𝗆[Q,Wi][fi|LT[L]]]C.\nu(\mathcal{L})\geqslant\mathop{\mathbb{E}}_{W_{i}\in\mathcal{V}}\left[\Pr_{L\in{\sf Zoom}[Q,W_{i}]}[f_{i}|_{L}\equiv T[L]]\right]\geqslant C.

By Lemma 5.14 with a=r1,r=r2,a=r_{1},r=r_{2}, and j=2j=2\ell, it follows that

μ()12(C3qr2(2r1)mqr1)C5.\mu(\mathcal{L})\geqslant\frac{1}{2}\cdot\left(C-\frac{3q^{r_{2}}(2\ell-r_{1})}{\sqrt{m}q^{-r_{1}}}\right)\geqslant\frac{C}{5}.

Summing everything up, this shows that there is a zoom-out function pair (V,h)(V^{\star},h^{\star}) such that VWiV^{\star}\supsetneq W_{i} and h|Wifih^{\star}|_{W_{i}}\equiv f_{i} for at least one ii, and PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[h|LT[L]|QLV]C5\Pr_{L\subseteq{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[h^{\star}|_{L}\equiv T[L]\;|\;Q\subseteq L\subseteq V^{\star}]\geqslant\frac{C}{5}. This contradicts the assumption that (Wi,fi)(W_{i},f_{i}) is (C,15)(C,\frac{1}{5})-maximal with respect to TT on QQ. ∎

5.4 An Auxiliary Lemma

We conclude this section with an auxiliary lemma that will be used in the analysis.

Lemma 5.22.

Let UU be a fixed question to the first prover in the Outer PCP consisting of 3k3k-variables in some set of kk equations. Let VUV\subseteq U be a random question to the second prover chosen according to the Outer PCP. Let W𝔽qUW\subseteq\mathbb{F}_{q}^{U} be a subspace of co-dimension ss. Then, with probability at least 12sβ1-2s\beta over the choice of the question VV, we have

dim(W𝔽qV)=|V|s.\dim(W\cap\mathbb{F}_{q}^{V})=|V|-s.
Proof.

Say that WW is given by the constraints h1,x=0,,hs,x=0\langle h_{1},x\rangle=0,\ldots,\langle h_{s},x\rangle=0 for h1,,hsh_{1},\ldots,h_{s} linearly independent. We can view 𝔽qV𝔽qU\mathbb{F}_{q}^{V}\subseteq\mathbb{F}_{q}^{U} as being defined by the constraints vi1,x=0,,vit,x=0\langle v_{i_{1}},x\rangle=0,\ldots,\langle v_{i_{t}},x\rangle=0, where i1,,iti_{1},\ldots,i_{t} correspond to the variables xi1,,xitx_{i_{1}},\ldots,x_{i_{t}} in UVU\setminus V. The event dim(W𝔽qV)=|V|s\dim(W\cap\mathbb{F}_{q}^{V})=|V|-s is equivalent to the event that h1,,hb,vi1,,vith_{1},\ldots,h_{b},v_{i_{1}},\ldots,v_{i_{t}} are linearly independent. Since h1,,hsh_{1},\ldots,h_{s} are linearly independent, there are ss-coordinates such that the restrictions of h1,,hsh_{1},\ldots,h_{s} to these ss-coordinates are linearly independent. If none of i1,,iti_{1},\ldots,i_{t} are in this set of ss coordinates, then h1,,hb,xi1,,xith_{1},\ldots,h_{b},x_{i_{1}},\ldots,x_{i_{t}} are linearly independent. Conditioned on the size of UVU\setminus V being tt, this event happens with probability at least 1ts/k1-ts/k, and as the expectation of tt is 2βk2\beta k it follows that the probability in question is at least 12sβ1-2s\beta. ∎

6 Analysis of the PCP

In this section we show completeness and soundness analysis of the composed PCP construction Ψ\Psi from Section 4. As usual, the completeness analysis is straightforward and the soundness analysis will consist the bulk of our effort.

6.1 Completeness

Suppose that the 𝟥𝖫𝗂𝗇{\sf 3Lin} instance (X,𝖤𝗊)(X,{\sf Eq}) has an assignment σ:X𝔽q\sigma:X\xrightarrow[]{}\mathbb{F}_{q} that satisfies at least 1ε11-\varepsilon_{1} of the equations in 𝖤𝗊{\sf Eq}. Let 𝒰𝗌𝖺𝗍𝒰\mathcal{U}_{{\sf sat}}\subseteq\mathcal{U} be the set of all U=(e1,,ek)U=(e_{1},\ldots,e_{k}) where all kk equations e1,,eke_{1},\ldots,e_{k} are satisfied. Then, |𝒰𝗌𝖺𝗍|(1kε1)𝒰|\mathcal{U}_{{\sf sat}}|\geqslant(1-k\varepsilon_{1})\mathcal{U}. We identify σ\sigma with the linear function from 𝔽qX𝔽q\mathbb{F}_{q}^{X}\to\mathbb{F}_{q}, assigning the value σ(i)\sigma(i) to the iith elementary basis element eie_{i}. Abusing notation, we denote this linear map by σ\sigma as well.

For each U𝒰𝗌𝖺𝗍U\in\mathcal{U}_{{\sf sat}} and vertex LHUL\oplus H_{U}, we set T1[LHU]σ|LHUT_{1}[L\oplus H_{U}]\equiv\sigma|_{L\oplus H_{U}}. Since U𝒰𝗌𝖺𝗍U\in\mathcal{U}_{{\sf sat}}, these assignments satisfy the side conditions. For all other UU’s, set T1[LHU]T_{1}[L\oplus H_{U}] so that the side conditions of HUH_{U} are satisfied and T1[LHU]|Lσ|LT_{1}[L\oplus H_{U}]|_{L}\equiv\sigma|_{L}. Such an assignment is possible because LHU={0}L\cap H_{U}=\{0\}. Similarly, the table T2T_{2} is defined as T2[R]σ|RT_{2}[R]\equiv\sigma|_{R}.

Sampling a constraint, note that the constraint is satisfied whenever the LHUL^{\prime}\oplus H_{U^{\prime}} chosen in step 33 of the test satisfies that U𝒰𝗌𝖺𝗍U^{\prime}\in\mathcal{U}_{{\sf sat}}. As the marginal distribution of LHUL^{\prime}\oplus H_{U^{\prime}} is uniform, 888This is true because first a clique is chosen with probability that is proportional to its size and then a vertex is sampled uniformly from the clique. the distribution of UU^{\prime} is uniform. It follows that the constraint is satisfied whenever U𝒰𝗌𝖺𝗍U^{\prime}\in\mathcal{U}_{{\sf sat}}, which happens with probability at least 1kε11-k\varepsilon_{1}. Thus, 𝗏𝖺𝗅(Ψ)1kε1{\sf val}(\Psi)\geqslant 1-k\varepsilon_{1}.

6.2 Soundness

In this section we relate the soundness of the composed PCP to that of the outer PCP and prove Lemma 6.1. More precisely, we show:

Lemma 6.1.

For all δ>0\delta>0 there are rr\in\mathbb{N} and c(δ)>0c(\delta)>0 such that the following holds. Let Gβ,rkG_{\beta,r}^{\otimes k} parallel repetition of the Smooth Variable versus Equation Game with advice described in Section 3.1.4, and let Ψ\Psi be composed PCP described in Section 4. If 𝗏𝖺𝗅(Gβ,rk)<qc(δ)O(2)\operatorname{{\sf val}}(G_{\beta,r}^{\otimes k})<q^{-c(\delta)\cdot O(\ell^{2})}, then 𝗏𝖺𝗅(Ψ)q2(11000δ)\operatorname{{\sf val}}(\Psi)\leqslant q^{-2(1-1000\delta)\ell}.

The rest of Section 6 is devoted to the proof of Lemma 6.1. The proof heavily relies on the tools from Section 5. Assume, as in lemma Lemma 6.1, that the 𝗏𝖺𝗅(Gβ,rk)<qc(δ)O()2\operatorname{{\sf val}}(G_{\beta,r}^{\otimes k})<q^{-c(\delta)\cdot O(\ell)^{2}}, and suppose for the sake of contradiction that there are tables T1T_{1} and T2T_{2} that are ε\varepsilon-consistent for εq2(11000δ)\varepsilon\geqslant q^{-2\ell(1-1000\delta)}. To arrive at a contradiction, we show that this implies strategies for the two provers that with success probability greater than 𝗏𝖺𝗅(Gβ,rk)\operatorname{{\sf val}}(G_{\beta,r}^{\otimes k}).

6.2.1 Clique Consistency

To start, we will reduce to the case where T1T_{1} satisfies a condition called clique-consistency.

Definition 6.2.

We say an assignment TT to 𝒜\mathcal{A} is clique consistent if for every vertex L1HU1L_{1}\oplus H_{U_{1}} and for every L2HU2,L3HU3[L1HU1]L_{2}\oplus H_{U_{2}},L_{3}\oplus H_{U_{3}}\in[L_{1}\oplus H_{U_{1}}], the assignments T[L2HU2]T[L_{2}\oplus H_{U_{2}}] and T[L3HU3]T[L_{3}\oplus H_{U_{3}}] satisfy the 11-to-11 constraint between L2HU2L_{2}\oplus H_{U_{2}} and L3HU3L_{3}\oplus H_{U_{3}} as specified in Lemma 4.2.

The following lemma shows that if T1T_{1} and T2T_{2} are ε\varepsilon-consistent assignments to Ψ\Psi, then there are clique-consistent assignments T~1\tilde{T}_{1} and T~2\tilde{T}_{2} that are also ε\varepsilon-consistent.

Lemma 6.3.

Suppose that the assignments T1T_{1} and T2T_{2} are ε\varepsilon-consistent, then there is a clique-consistent assignment T~1\tilde{T}_{1} such that T~1\tilde{T}_{1} and T2T_{2} are ε\varepsilon-consistent.

Proof.

Partition 𝒜\mathcal{A} into cliques, 𝒜=Clique1Cliquem\mathcal{A}=\textsf{Clique}_{1}\sqcup\cdots\sqcup\textsf{Clique}_{m}. For each ii, choose a random LHUCliqueiL\oplus H_{U}\in\textsf{Clique}_{i} uniformly, and for every LHUCliqueiL^{\prime}\oplus H_{U^{\prime}}\in\textsf{Clique}_{i} assign T~1[LHU]\tilde{T}_{1}[L^{\prime}\oplus H_{U^{\prime}}] in the unique way that is consistent with T1[LHU]T_{1}[L\oplus H_{U}] and the side conditions of UU^{\prime} as described in Lemma 4.2. It is clear that T~1\tilde{T}_{1} is clique consistent, and we next analyze the expected fraction of constraints that T~1\tilde{T}_{1} and T2T_{2} satisfy.

Note that an alternative description of sampling a constraint in Ψ\Psi proceeds as follows. First choose a clique Cliquei\textsf{Clique}_{i} with probability that is proportional to by its size, and then choose LHUCliqueiL\oplus H_{U}\in\textsf{Clique}_{i} in the first step. The rest of the sampling procedure is the same. Let P(LHU)P(L\oplus H_{U}) be the probability that the test passes conditioned on LHUL\oplus H_{U} being chosen in the second step. It is clear that every vertex in the clique has equal probability of being chosen, therefore the probability of passing if Cliquei\textsf{Clique}_{i} chosen is

1|Cliquei|LHUCliqueiP(LHU).\frac{1}{|\textsf{Clique}_{i}|}\sum_{L\oplus H_{U}\in\textsf{Clique}_{i}}P(L\oplus H_{U}).

On the other hand, the expected fraction of constraints satisfied by T~1\tilde{T}_{1} and T2T_{2} (over the randomness of choosing T~1\tilde{T}_{1}) is

LHU1|Cliquei|P(LHU)=1|Cliquei|LHUCliqueiP(LHU).\sum_{L\oplus H_{U}}\frac{1}{|\textsf{Clique}_{i}|}\cdot P(L\oplus H_{U})=\frac{1}{|\textsf{Clique}_{i}|}\sum_{L\oplus H_{U}\in\textsf{Clique}_{i}}P(L\oplus H_{U}).

To see this, note that for any LHUL\oplus H_{U}, 1|Cliquei|\frac{1}{|\textsf{Clique}_{i}|} is the probability that T1[LHU]T_{1}[L\oplus H_{U}] is used to define T~1\tilde{T}_{1} on Cliquei\textsf{Clique}_{i}. If this is the case, then the probability the test passes on T~1\tilde{T}_{1} within Cliquei\textsf{Clique}_{i} is P(LHU)P(L\oplus H_{U}).

Since this holds over every clique, it follows that the expected fraction of constraints satisfied by T~1\tilde{T}_{1} equals the fraction of constraints satisfied by T1T_{1} and T2T_{2}. In particular, there is a choice of T~1\tilde{T}_{1} such that together with T2T_{2} it satisfies at least ε\varepsilon fraction of the constraints. ∎

Applying Lemma 6.3 we conclude that there are clique-consistent assignments to Ψ\Psi that are ε\varepsilon-consistent, and henceforth we assume that T1T_{1} are clique-consistent to begin with. We remark that, in the notation of Section 4.1.4, the benefit of having a clique-consistent assignment is that the constraint that the verifier checks is equivalent to checking that T1[LHU]|RT2[R]T_{1}[L\oplus H_{U}]|_{R}\equiv T_{2}[R]. The latter check is a test which in performed within the space 𝔽qU\mathbb{F}_{q}^{U} of the first prover. We will use this fact in the next section.

6.2.2 A Strategy for the First Prover

Let p(U)p(U) be the consistency of T1T_{1} and T2T_{2} conditioned on UU being the question to the first prover. As we are assuming that the overall success probability is at least ε\varepsilon, EU[p(U)]εE_{U}[p(U)]\geqslant\varepsilon. By an averaging argument, p(U)ε2p(U)\geqslant\frac{\varepsilon}{2} for at least ε2\frac{\varepsilon}{2}-fraction of the UU’s. Call such UU’s good and let 𝒰𝗀𝗈𝗈𝖽\mathcal{U}_{{\sf good}} be the set of good UU’s.

Let U𝒰U\in\mathcal{U} be the question to the first prover and let QQ be the advice. If U𝒰𝗀𝗈𝗈𝖽U\notin\mathcal{U}_{{\sf good}}, then the first prover gives up, so henceforth assume that U𝒰𝗀𝗈𝗈𝖽U\in\mathcal{U}_{{\sf good}}. For such UU, the test of the inner PCP passes with probability at least ε2\frac{\varepsilon}{2}. More concretely, we have

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}dim(R)=2(1δ)[T1[L]|RT2[R]|RL]ε2.\Pr_{\begin{subarray}{c}L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}\\ \dim(R)=2(1-\delta)\ell\end{subarray}}[T_{1}[L]|_{R}\equiv T_{2}[R]\;|\;R\subseteq L]\geqslant\frac{\varepsilon}{2}.

Next, the first prover chooses an integer 0r10δ0\leqslant r\leqslant\frac{10}{\delta} uniformly, and takes QQ to be the span of the first rr-advice vectors. By Theorem 5.3, there are r1,r2r_{1},r_{2} satisfying r1+r210δr_{1}+r_{2}\leqslant\frac{10}{\delta} such that for at least q62q^{-6\ell^{2}} of the Q𝔽qUQ\subseteq\mathbb{F}_{q}^{U} of dimension r1r_{1}, there exists WQ𝔽qUW_{Q}\subseteq\mathbb{F}_{q}^{U} containing QHUQ\oplus H_{U} of codimension r210δr_{2}\leqslant\frac{10}{\delta} and a linear function gQ,WQ:WQ𝔽qg_{Q,W_{Q}}:W_{Q}\xrightarrow[]{}\mathbb{F}_{q} that satisfies

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}[gQ,WQ|LHUT1[LHU]|QLWQ]q2(11000δ2)5.\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}}[g_{Q,W_{Q}}|_{L\oplus H_{U}}\equiv T_{1}[L\oplus H_{U}]\;|\;Q\subseteq L\subseteq W_{Q}]\geqslant\frac{q^{-2(1-1000\delta^{2})\ell}}{5}. (9)

For simplicity, set C=q2(11000δ2)5C=\frac{q^{-2(1-1000\delta^{2})\ell}}{5}. With probability at least δ/10\delta/10, they choose r=r1r=r_{1}, where r1r_{1} is the parameter from Theorem 5.3. Call these dimension r1r_{1} subspaces QQ lucky and let 𝒬𝗅𝗎𝖼𝗄𝗒\mathcal{Q}_{{\sf lucky}} be the set of all lucky Q𝔽qUQ\subseteq\mathbb{F}_{q}^{U}. For our analysis, we only analyze the case where the first prover chooses chooses r=r1r=r_{1}, which occurs with probability at least δ20\frac{\delta}{20}

For each QQ such that Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}} and QHU={0}Q\cap H_{U}=\{0\}, the first prover chooses a WQW_{Q} of codimension at most 10δ\frac{10}{\delta} and linear function gQ,WQ:WQ𝔽qg_{Q,W_{Q}}:W_{Q}\xrightarrow[]{}\mathbb{F}_{q} that satisfies the side conditions on HUH_{U} and Equation (9). For such QQ that are in 𝒬𝗅𝗎𝖼𝗄𝗒\mathcal{Q}_{{\sf lucky}} and satisfy QHU={0}Q\cap H_{U}=\{0\}, define

Q={L𝖹𝗈𝗈𝗆[Q,WQ]|gQ,WQ|LT1[LHU]|L}.\mathcal{L}_{Q}=\{L\in{\sf Zoom}[Q,W_{Q}]\;|\;g_{Q,W_{Q}}|_{L}\equiv T_{1}[L\oplus H_{U}]|_{L}\}.

For QQ such that Q𝒬𝗅𝗎𝖼𝗄𝗒Q\notin\mathcal{Q}_{{\sf lucky}} or QHU{0}Q\cap H_{U}\neq\{0\} define Q=\mathcal{L}_{Q}=\emptyset. Finally, define

={(x1,,x2)(𝔽qU)2|span(x1,,xr1)𝒬𝗅𝗎𝖼𝗄𝗒,span(x1,,x2)span(x1,,xr1)},\mathcal{L}=\{(x_{1},\ldots,x_{2\ell})\in\left(\mathbb{F}_{q}^{U}\right)^{2\ell}\;|\;\operatorname{span}(x_{1},\ldots,x_{r_{1}})\in\mathcal{Q}_{{\sf lucky}},\operatorname{span}(x_{1},\ldots,x_{2\ell})\in\mathcal{L}_{\operatorname{span}(x_{1},\ldots,x_{r_{1}})}\},

and let 𝒬𝗌𝗆𝗈𝗈𝗍𝗁\mathcal{Q}_{{\sf smooth}} denote set of r1r_{1}-dimensional subspaces QQ such that

𝒟Q()0.8𝒟Q()η20,\mathcal{D}^{\prime}_{Q}(\mathcal{L})\geqslant 0.8\mathcal{D}_{Q}(\mathcal{L})-\eta^{20}, (10)

where η=q100\eta=q^{-\ell^{100}}. By definition of 𝒟Q\mathcal{D}_{Q} and \mathcal{L}, if Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, we have

𝒟Q()\displaystyle\mathcal{D}_{Q}(\mathcal{L}) =Prx=(x1,,x2)𝔽qU×2[gQ,WQ|span(x)T1[span(x)HU]|span(x)|spanr1(x)=Q]\displaystyle=\Pr_{x=(x_{1},\ldots,x_{2\ell})\in\mathbb{F}_{q}^{U\times 2\ell}}[g_{Q,W_{Q}}|_{\operatorname{span}(x)}\equiv T_{1}[\operatorname{span}(x)\oplus H_{U}]|_{\operatorname{span}(x)}\;|\;\operatorname{span}_{r_{1}}(x)=Q]
=Prx=(x1,,x2)𝔽qU×2[gQ,WQ|span(x)T1[span(xHU)|span(x)]span(x)WQ|spanr1(x)=Q]\displaystyle=\Pr_{x=(x_{1},\ldots,x_{2\ell})\in\mathbb{F}_{q}^{U\times 2\ell}}[g_{Q,W_{Q}}|_{\operatorname{span}(x)}\equiv T_{1}[\operatorname{span}(x\oplus H_{U})|_{\operatorname{span}(x)}]\land\operatorname{span}(x)\subseteq W_{Q}\;|\;\operatorname{span}_{r_{1}}(x)=Q]
=PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}[gQ,WQ|LHUT1[LHU]LWQ|QL]\displaystyle=\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}}[g_{Q,W_{Q}}|_{L\oplus H_{U}}\equiv T_{1}[L\oplus H_{U}]\land L\subseteq W_{Q}\;|\;Q\subseteq L]

where the second transition is because, by definition, every LQL\in\mathcal{L}_{Q} is contained in WQW_{Q}. For the third transition, we are ignoring the probability that the xix_{i}’s are linearly dependent and span(x)HU{0}\operatorname{span}(x)\cap H_{U}\neq\{0\}. Indeed, the probability that either of these events occur is at most q23k+q22kq^{2\ell-3k}+q^{2\ell-2k}, and is negligible anyways.

Continuing, for Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, we have

𝒟Q()\displaystyle\mathcal{D}_{Q}(\mathcal{L}) =PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}[gQ,WQ|LHUT1[LHU]LWQ|QL]\displaystyle=\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}}[g_{Q,W_{Q}}|_{L\oplus H_{U}}\equiv T_{1}[L\oplus H_{U}]\land L\subseteq W_{Q}\;|\;Q\subseteq L]
=PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}[gQ,WQ|LHUT1[LHU]|QLWQ]\displaystyle=\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}}[g_{Q,W_{Q}}|_{L\oplus H_{U}}\equiv T_{1}[L\oplus H_{U}]\;|\;Q\subseteq L\subseteq W_{Q}]
PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}[LWQ|QL]\displaystyle\cdot\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}}[L\subseteq W_{Q}\;|\;Q\subseteq L]
qr2(2r1)C,\displaystyle\geqslant q^{-r_{2}(2\ell-r_{1})}\cdot C, (11)

where in the third transition uses Equation (9) to lower bound the first term by CC.

If either Q𝒬𝗅𝗎𝖼𝗄𝗒Q\notin\mathcal{Q}_{{\sf lucky}}, QHU{0}Q\cap H_{U}\neq\{0\}, or Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\notin\mathcal{Q}_{{\sf smooth}}, then the first prover gives up. Otherwise the prover extends the function gQg_{Q} to a linear function on the entire space 𝔽qU\mathbb{F}_{q}^{U} randomly, and we denote this extension by g:𝔽qU𝔽qg:\mathbb{F}_{q}^{U}\xrightarrow[]{}\mathbb{F}_{q}. The prover outputs the string sQ,Us_{Q,U} as their answer where sQ,U𝔽qUs_{Q,U}\in\mathbb{F}_{q}^{U} is the unique string such that g(x)=sQ,U,xg(x)=\langle s_{Q,U},x\rangle for all x𝔽qUx\in\mathbb{F}_{q}^{U}. As gQ,WQg_{Q,W_{Q}}, and by extension gg, respects the side conditions, it follows that sQ,Us_{Q,U} satisfies the kk-linear equations of UU.

6.2.3 A Strategy for the Second Prover

Let VV be the question to the second prover. The second prover will use a table T~1\tilde{T}_{1} to derive their strategy. The table T~1\tilde{T}_{1} is obtained from T1T_{1} as follows. For a question VV to the second prover, let UVU\supseteq V be an arbitrary question to the first prover. For all 22\ell-dimensional subspaces L𝔽qVL\subseteq\mathbb{F}_{q}^{V}, define

T~1[L]T1[LHU]|L.\tilde{T}_{1}[L]\equiv T_{1}[L\oplus H_{U}]|_{L}.

In order to make sure that T~1\tilde{T}_{1} is well defined, we note two things. First, the subspace LHUL\oplus H_{U} can be viewed as a subspace of 𝔽qU\mathbb{F}_{q}^{U} because each L𝔽qVL\subseteq\mathbb{F}_{q}^{V} can be “lifted” to a subspace of 𝔽qU\mathbb{F}_{q}^{U} by inserting 0’s into the coordinates corresponding to VUV\setminus U. Second, note that the choice of UVU\supseteq V does not matter when defining T~1[L]\tilde{T}_{1}[L]. Indeed, for a fixed LL, the vertices LHUL\oplus H_{U} over all UVU\supseteq V are in the same clique. Since T1T_{1} is clique consistent, it does not matter which UU is chosen when defining T~1[L]\tilde{T}_{1}[L], as all choices lead to the same function T1[LHU]|LT_{1}[L\oplus H_{U}]|_{L}. Therefore the second prover can construct the table T~1\tilde{T}_{1}.

After constructing T~1\tilde{T}_{1}, the second prover then chooses a dimension 0r10δ0\leqslant r\leqslant\frac{10}{\delta} uniformly for the advice QQ. Note that with probability at least δ20\frac{\delta}{20} the second prover also chooses r=r1r=r_{1}. The second prover then chooses a zoom-out function pair (W𝗌𝖾𝖼𝗈𝗇𝖽,gQ,W𝗌𝖾𝖼𝗈𝗇𝖽)(W_{{\sf second}},g_{Q,W_{{\sf second}}}) that is

(C45r2,15)\left(\frac{C}{4\cdot 5^{r_{2}}},\frac{1}{5}\right)

-maximal with respect to T~1\tilde{T}_{1} on QQ if one exists (and gives up otherwise).

Finally, the second prover extends the function gQ,W𝗌𝖾𝖼𝗈𝗇𝖽g_{Q,W_{{\sf second}}} randomly to a linear function on 𝔽qV\mathbb{F}_{q}^{V} to arrive at their answer. The resulting function is linear and it is equal to the inner product function ysQ,V,yy\xrightarrow[]{}\langle s_{Q,V},y\rangle for some unique string sQ,V𝔽qVs_{Q,V}\in\mathbb{F}_{q}^{V}. The second prover outputs sQ,Vs_{Q,V} as their answer.

6.2.4 The Success Probability of the Provers

In order to be successful, a series of events must occur. We go through each one and state the probability that each occurs. At the end this yields a lower bound on the provers’ success probability. We remark that the analysis of this sections requires Lemmas 5.4 and  5.6, so recall that kk and β\beta are set according to Equation (4) in Section 1.2.6 so that these lemmas hold.

First, the provers need U𝒰𝗀𝗈𝗈𝖽U\in\mathcal{U}_{{\sf good}}, which occurs with probability at least ε2\frac{\varepsilon}{2}. Assuming that this occurs, the provers then both need to choose r=r1r=r_{1} for the dimension of their zoom-in, which happens with probability at least δ2101\frac{\delta^{2}}{101}. If both provers choose r=r1r=r_{1}, they both receive advice QQ as the span of r1r_{1} random vectors.

The provers then need Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, QHU={0}Q\cap H_{U}=\{0\}, and Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\in\mathcal{Q}_{{\sf smooth}}. When analyzing the probability that these three events occur, we need to recall that the advice vectors are actually drawn according to distribution 𝒟r1\mathcal{D}^{\prime}_{r_{1}}, the distribution described in Section 5.2.3. We will analyze the probability that the three events occur under 𝒟r1\mathcal{D}_{r_{1}} and then appeal to the covering property of Lemma 5.6. By Theorem 5.2, the first item occurs with probability at least q62q^{-6\ell^{2}}. On the other hand the probability that the second item does not occur is at most i=0r1qiqkq3kqr1+12k\sum_{i=0}^{r_{1}}\frac{q^{i}q^{k}}{q^{3k}}\leqslant q^{r_{1}+1-2k}, while the probability that the third item does not occur is at most η20\eta^{20} by Lemma 5.5. Altogether we get that with probability at least

q62qr+12kη20q72q^{-6\ell^{2}}-q^{r^{*}+1-2k}-\eta^{20}q^{-7\ell^{2}}

under 𝒟r1\mathcal{D}_{r_{1}}, we have Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, QHU={0}Q\cap H_{U}=\{0\}, and Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\in\mathcal{Q}_{{\sf smooth}}. By Lemma 5.6, we have that Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, QHU={0}Q\cap H_{U}=\{0\}, and Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\in\mathcal{Q}_{{\sf smooth}} with probability at least q82q^{-8\ell^{2}} under 𝒟r1\mathcal{D}_{r_{1}}^{\prime} - the distribution which the QQ is actually drawn from.

Now let us assume that U𝒰𝗀𝗈𝗈𝖽U\in\mathcal{U}_{{\sf good}}, and both provers receive an r1r_{1}-dimensional advice QQ such that Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, QHU={0}Q\cap H_{U}=\{0\}, and Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\in\mathcal{Q}_{{\sf smooth}}. The first prover chooses the function gQ:WQ𝔽qg_{Q}:W_{Q}\xrightarrow[]{}\mathbb{F}_{q}. Write codim(WQ)=r2\operatorname{codim}(W_{Q})=r_{2}. Since Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, by Equation (6.2.2) we have

𝒟Q()qr2(2r1)C,\mathcal{D}_{Q}(\mathcal{L})\geqslant q^{-r_{2}(2\ell-r_{1})}\cdot C,

since by definition \mathcal{L} contains at least CC-fraction of L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W], which is in turn at least qr2(2r1)q^{-r_{2}(2\ell-r_{1})} fraction of L𝖹𝗈𝗈𝗆[Q,𝔽qU]L\in{\sf Zoom}[Q,\mathbb{F}_{q}^{U}]. Next because Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\in\mathcal{Q}_{{\sf smooth}}, we have

𝒟Q()0.8𝒟Q()η60qr2(2r1)C2,\mathcal{D}^{\prime}_{Q}(\mathcal{L})\geqslant 0.8\cdot\mathcal{D}_{Q}(\mathcal{L})-\eta^{60}\geqslant q^{-r_{2}(2\ell-r_{1})}\cdot\frac{C}{2}, (12)

by Equation (10).

Now let WQ[V]=WQ𝔽qVW_{Q}[V]=W_{Q}\cap\mathbb{F}_{q}^{V}. By Lemma 5.22, with probability at least 12βr21-2\beta r_{2} we have that codim(WQ[V])=r1\operatorname{codim}(W_{Q}[V])=r_{1} inside of 𝔽qV\mathbb{F}_{q}^{V}. Combining this with an averaging argument on Equation (12), we have that with probability at least C52βr2C6\frac{C}{5}-2\beta r_{2}\geqslant\frac{C}{6} over VV,

Prxi𝔽qV,wiHU[span(x1+w1,,x2+w2)|spanr1(x)=Q]=𝒟Q()qr2(2r1)C4,\begin{split}\Pr_{x_{i}\in\mathbb{F}_{q}^{V},w_{i}\in H_{U}}[\operatorname{span}(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell})\in\mathcal{L}\;|\;\operatorname{span}_{r_{1}}(x)=Q]&=\mathcal{D}^{\prime}_{Q}(\mathcal{L})\\ &\geqslant q^{-r_{2}(2\ell-r_{1})}\cdot\frac{C}{4},\end{split} (13)

and codim(WQ[V])=r1\operatorname{codim}(W_{Q}[V])=r_{1}. We call such VV consistent. In the probability above, and henceforth, we view vectors xi𝔽qVx_{i}\in\mathbb{F}_{q}^{V} as vectors in 𝔽qU\mathbb{F}_{q}^{U} with 0’s appended in the missing coordinates. At this point there is the slight issue that span(x1+w1,,x2+w2)\operatorname{span}(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell}) does not actually correspond to a random entry of the second prover’s table, T~1\tilde{T}_{1}. Indeed, the second prover can only choose (x1,,x2)𝔽qV(x_{1},\ldots,x_{2\ell})\in\mathbb{F}_{q}^{V}, choose some question UU to the first prover that contains VV, lift these to 𝔽qU\mathbb{F}_{q}^{U} by inserting zeros in the missing coordinates, and look at the entry LHUL\oplus H_{U} where L=span(x1,,x2)L=\operatorname{span}(x_{1},\ldots,x_{2\ell}). They do not know the question UU and the side conditions HUH_{U}, and hence could not sample the wiHUw_{i}\in H_{U}. However, notice that for any w1,,w2HUw_{1},\ldots,w_{2\ell}\in H_{U}, we have that,

span(x1,,x2)HU=span(x1+w1,,x2+w2)HU.\operatorname{span}(x_{1},\ldots,x_{2\ell})\oplus H_{U}=\operatorname{span}(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell})\oplus H_{U}.

We can thus view the w1,,w2HUw_{1},\ldots,w_{2\ell}\in H_{U} being sampled and added to x1,,x2x_{1},\ldots,x_{2\ell} as a virtual step. In the next two equations, we ignore the probability that (x1,,x2)(x_{1},\ldots,x_{2\ell}) or (x1+w1,,x2+w2)(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell}) are not linearly independent or have spans intersecting HUH_{U} to make the expressions above simpler. This probability is at most q2+qk+2q3k\frac{q^{2\ell}+q^{k+2\ell}}{q^{3k}} and is negligible anyways. We have,

PrL𝖦𝗋𝖺𝗌𝗌q(𝔽qV,2)[T~1[L]=gQ|L|QLWQ[V]]=PrL=span(x1,,x2)[T~1[L]=gQ|L|QLWQ[V]]=PrL=span(x1,,x2)[T1[LHU]=gQ|L|QLWQ[V]]=PrL=span(x1+w1,,x2+w2)[T1[LHU]=gQ|L|QLWQ[V]]=PrL=span(x1+w1,,x2+w2)[T~1[L]=gQ|L|QLWQ[V]].\begin{split}&\Pr_{L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{V},2\ell)}[\tilde{T}_{1}[L]=g_{Q}|_{L}\;|\;Q\subseteq L\subseteq W_{Q}[V]]\\ &\qquad\qquad\qquad=\Pr_{L=\operatorname{span}(x_{1},\ldots,x_{2\ell})}[\tilde{T}_{1}[L]=g_{Q}|_{L}\;|\;Q\subseteq L\subseteq W_{Q}[V]]\\ &\qquad\qquad\qquad=\Pr_{L=\operatorname{span}(x_{1},\ldots,x_{2\ell})}[T_{1}[L\oplus H_{U}]=g_{Q}|_{L}\;|\;Q\subseteq L\subseteq W_{Q}[V]]\\ &\qquad\qquad\qquad=\Pr_{L=\operatorname{span}(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell})}[T_{1}[L\oplus H_{U}]=g_{Q}|_{L}\;|\;Q\subseteq L\subseteq W_{Q}[V]]\\ &\qquad\qquad\qquad=\Pr_{L=\operatorname{span}(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell})}[\tilde{T}_{1}[L]=g_{Q}|_{L}\;|\;Q\subseteq L\subseteq W_{Q}[V]].\end{split}

This last probability can be related to 𝒟Q()\mathcal{D}^{\prime}_{Q}(\mathcal{L})

PrL=span(x1+w1,,x2+w2)[T~1[L]=gQ|L|QLWQ[V]]qr2(2r1)Prxi𝔽qV,wiHU[(x1+w1,,x2+w2)|spanr1((xi+wi))=Q]=qr2(2r1)𝒟Q()C4.\begin{split}&\Pr_{L=\operatorname{span}(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell})}[\tilde{T}_{1}[L]=g_{Q}|_{L}\;|\;Q\subseteq L\subseteq W_{Q}[V]]\\ &\geqslant q^{r_{2}(2\ell-r_{1})}\cdot\Pr_{x_{i}\in\mathbb{F}_{q}^{V},w_{i}\in H_{U}}[(x_{1}+w_{1},\ldots,x_{2\ell}+w_{2\ell})\in\mathcal{L}\;|\;\operatorname{span}_{r_{1}}((x_{i}+w_{i}))=Q]\\ &=q^{r_{2}(2\ell-r_{1})}\cdot\mathcal{D}^{\prime}_{Q}(\mathcal{L})\\ &\geqslant\frac{C}{4}.\end{split}

By Lemma 5.18, there exists some (WQ[V],gQ,WQ[V])(W^{\prime}_{Q}[V],g_{Q,W^{\prime}_{Q}[V]}) that is (C45r2,15)\left(\frac{C}{4\cdot 5^{r_{2}}},\frac{1}{5}\right)-maximal and satisfies WQ[V]WQ[V]W^{\prime}_{Q}[V]\supseteq W_{Q}[V], gQ,WQ[V]:WQ[V]𝔽qg_{Q,W^{\prime}_{Q}[V]}:W^{\prime}_{Q}[V]\xrightarrow[]{}\mathbb{F}_{q} is linear, and gQ,WQ[V]|W[V]=gQ|WQ[V]g_{Q,W^{\prime}_{Q}[V]}|_{W[V]}=g_{Q}|_{W_{Q}[V]}. By Theorem 5.21, the number of (C45r2,15)\left(\frac{C}{4\cdot 5^{r_{2}}},\frac{1}{5}\right) is at most

M=40δ52r2+2C2q100(t1)!r2ξ1qc(δ),M=\frac{40}{\delta}\cdot 5^{2r_{2}+2}C^{-2}\cdot q^{100\left(t-1\right)!r^{2}\ell\xi^{-1}}\leqslant q^{c(\delta)\ell},

where c(δ)c(\delta) is some function depending only on δ\delta. Thus, the second prover chooses (WQ[V],gQ,WQ[V])(W^{\prime}_{Q}[V],g_{Q,W^{\prime}_{Q}[V]}) with probability at least 1M\frac{1}{M}. Finally, if the second prover chooses (WQ[V],gQ,WQ[V])(W^{\prime}_{Q}[V],g_{Q,W^{\prime}_{Q}[V]}), then the provers succeed if both provers extend their functions, gQ|W[V]g_{Q}|_{W[V]} and gQ,WQ[V]g_{Q,W^{\prime}_{Q}[V]} in the same manner. This occurs with probability at least qcodim(W[V])q10/δq^{-\operatorname{codim}(W[V])}\geqslant q^{-10/\delta}.

Putting everything together, we get that the provers succeed with probability at least

ε2δ2101q82C51Mq10/δ=qc(δ)O(2),\frac{\varepsilon}{2}\cdot\frac{\delta^{2}}{101}\cdot q^{-8\ell^{2}}\cdot\frac{C}{5}\cdot\frac{1}{M}\cdot q^{-10/\delta}=q^{-c(\delta)\cdot O(\ell^{2})},

where the first term is the probability that U𝒰𝗀𝗈𝗈𝖽U\in\mathcal{U}_{{\sf good}}, the second term is the probability that both provers choose the same zoom-in dimension, the third term is the probability that Q𝒬𝗅𝗎𝖼𝗄𝗒Q\in\mathcal{Q}_{{\sf lucky}}, QHU={0}Q\cap H_{U}=\{0\}, Q𝒬𝗌𝗆𝗈𝗈𝗍𝗁Q\in\mathcal{Q}_{{\sf smooth}}, the fourth term is the probability that VV is consistent, the fifth term is the probability that the second prover chooses the a function that extends the first prover’s answer, and the final term is the probability that both provers extend their functions in the same manner.

This proves Lemma 6.1.

7 Proofs of the Main Theorems

7.1 Proof of Theorem 1.3

Theorem 1.3 follows by applying out PCP construction from Section 4 starting with an instance of 33-Lin from Theorem 2.1. We may take q=2q=2, fix δ>0\delta>0 to be a small constant and then take \ell sufficiently large compared to δ1\delta^{-1}, then kk and β\beta according to Equation (4), and finally take the completeness of the 33-Lin instance, 1η1-\eta, so that η<1/k\eta<1/k. It follows that if the original 33-Lin instance is at least 1η1-\eta satisfiable, then 𝗏𝖺𝗅(Ψ)1kη\operatorname{{\sf val}}(\Psi)\geqslant 1-k\eta. On the other hand, if the if the original instance is at most ss satisfiable for some constant s>0s>0, then by Claim 3.2, the value of the outer PCP is at most

𝗏𝖺𝗅(Gβ,rk)2Ω((1s)2qr+2c3)<qc(δ)O()2,\operatorname{{\sf val}}(G_{\beta,r}^{\otimes k})\leqslant 2^{-\Omega\left((1-s)^{2}q^{-r+\frac{2c}{3}\ell}\right)}<q^{-c(\delta)O(\ell)^{2}},

since we take \ell sufficiently large compared to δ1\delta^{-1}. By Lemma 6.1, it follows that if the original instance is at most ss satisfiable, then 𝗏𝖺𝗅(Ψ)q2(11000δ){\sf val}(\Psi)\leqslant q^{-2(1-1000\delta)\ell}. The proof is concluded as the alphabet size of Ψ\Psi is O(q2)O(q^{2\ell}).

7.2 Proof of Theorem 1.4

To show quasi-NP-hardness for approximate Quadratic Programming, we rely on the following result due to [ABH+05], who show a reduction from 22-Prover-11-Round Games to Quadratic Programming.

Theorem 7.1.

There is a reduction from a 2-Prover-1-Round Games, Ψ\Psi with graph G=(LR,E)G=(L\cup R,E) and alphabets ΣL,ΣR\Sigma_{L},\Sigma_{R} to a Quadratic Programming instance AA such that:

  • The running time of the reduction and the number of variables in AA is polynomial in |L|+|R||L|+|R| and 2|ΣL|2^{|\Sigma_{L}|}.

  • If 𝗏𝖺𝗅(Ψ)1η{\sf val}(\Psi)\geqslant 1-\eta, then 𝖮𝖯𝖳(A)1η1|L|+|R|{\sf OPT}(A)\geqslant 1-\eta-\frac{1}{|L|+|R|}.

  • If 𝗏𝖺𝗅(Ψ)ε{\sf val}(\Psi)\leqslant\varepsilon, then 𝖮𝖯𝖳(A)O(ε){\sf OPT}(A)\leqslant O(\varepsilon).

We are now ready to prove Theorem 1.4.

Proof of Theorem 1.4.

Starting with a 𝖲𝖠𝖳{\sf SAT} instance of size nn, which may be arbitrarily large, we take the instance of 𝖦𝖺𝗉𝟥𝖫𝗂𝗇{\sf Gap3Lin} from Theorem 2.2 of size N2O(log2n)N\leqslant 2^{O(\log^{2}n)} and field size q=Θ(logn)q=\Theta(\log n) which is a power of 22 as the starting point of our reduction. Take δ>0\delta>0 to be a small constant, and \ell to be a sufficiently large constant relative to δ1\delta^{-1} in our composed PCP. Finally, we pick kk and β\beta by Equation (4). This yields a 2O(klog2n)2^{O(k\log^{2}n)}-time reduction from 𝖲𝖠𝖳{\sf SAT} to a 2-Prover-1-Game on G=(LR,E)G=(L\cup R,E), with alphabets ΣL,ΣR\Sigma_{L},\Sigma_{R} and the following properties:

  • |R|+|L|=O(Nkq3k+2)|R|+|L|=O(N^{k}\cdot q^{3k+2\ell}).

  • |ΣR||ΣL|=q2|\Sigma_{R}|\leqslant|\Sigma_{L}|=q^{2\ell}.

  • The completeness is at least 1kη1-k\eta, where η=2Θ(logn)\eta=2^{-\Theta(\sqrt{\log n})}.

  • The soundness is at most q2(11000δ)q^{-2(1-1000\delta)\ell}.

Indeed, the first 33 properties are clear. For the soundness, as the original 33-Lin instance is at most 1ε1-\varepsilon satisfiable for ε=1/log3N\varepsilon=1/\log^{3}N, we get from Claim 3.2 that

𝗏𝖺𝗅(Gβ,rk)2Ω(ε2qr+2c3)qc(δ)O()2,\operatorname{{\sf val}}(G_{\beta,r}^{\otimes k})\leqslant 2^{-\Omega(\varepsilon^{-2}q^{-r+\frac{2c\ell}{3}})}\leqslant q^{-c(\delta)O(\ell)^{2}},

as \ell is sufficiently large relative to δ1\delta^{-1}, so the soundness of the composed PCP follows by Lemma 6.1. Applying the reduction of Theorem 7.1, we get a reduction to a Quadratic Programming instance AA such that,

  • The running time of the reduction and number of variables in AA are both

    M=poly(2O(log2n)q2(1+c)(logn)O(q2(1+c))2q2).M=\operatorname{poly}(2^{O(\log^{2}n)q^{2(1+c)\ell}}(\log n)^{O(q^{2(1+c)\ell})}2^{q^{2\ell}}).
  • If the original SAT instance is satisfiable, then

    𝖮𝖯𝖳12Ω(logn).{\sf OPT}\geqslant 1-2^{-\Omega(\sqrt{\log n})}.
  • If the original SAT instance is not satisfiable, then

    𝖮𝖯𝖳O(q2(11000δ)).{\sf OPT}\leqslant O(q^{-2(1-1000\delta)\ell}).

Note that

log(M)=q2(1+c)O(log2n),\log(M)=q^{2(1+c)\ell}O(\log^{2}n),

whereas the gap between the satisfiable and unsatisfiable cases is Ω(q2(11000δ))=1log(M)1O(δ)\Omega\left(q^{-2(1-1000\delta)\ell}\right)=\frac{1}{\log(M)^{1-O(\delta)}}. Altogether, this shows that for all ε>0\varepsilon>0 there is C>0C>0 such that unless 𝖭𝖯𝖣𝖳𝖨𝖬𝖤(2log(n)C){\sf NP}\subseteq{\sf DTIME}\left(2^{\log(n)^{C}}\right), there is no log(M)1ε\log(M)^{1-\varepsilon}-approximation algorithm for Quadratic Programming on MM variables. ∎

7.3 Proof of Theorem 1.5

In this section we prove Theorem 1.5, and for that we must first establish a version of Theorem 1.3 for bi-regular graphs of bounded degree. The proof of this requites minor modifications of our construction, as well as the right degree reduction technique of Moshkovitz and Raz [MR10].

7.3.1 Obtaining a Hard Instance of Bipartite Biregular 22-CSP

We first show that the 22-Prover-11-Round game from Theorem 1.3 can be transformed into a hard instance of biregular, bipartite 22-CSP with bounded degrees. This version may be useful for future applications, and is formally stated below. Call a bipartite 22-CSP (d1,d2)(d_{1},d_{2})-regular if the left degrees of its underlying graph are all d1d_{1}, and the right degrees of its underlying graph are all d2d_{2}.

Theorem 7.1.

For every φ,ε>0\varphi,\varepsilon>0, and sufficiently large RR\in\mathbb{N}, there exist d1,d2d_{1},d_{2}\in\mathbb{N} such that given a bipartite (d1,d2)(d_{1},d_{2})-regular 22-CSP, Ψ\Psi, with alphabet size RR, it is 𝖭𝖯{\sf NP}-hard to distinguish the following two cases:

  • Completeness: 𝗏𝖺𝗅(Ψ)1φ\operatorname{{\sf val}}(\Psi)\geqslant 1-\varphi,

  • Soundness: 𝗏𝖺𝗅(Ψ)1R1ε\operatorname{{\sf val}}(\Psi)\leqslant\frac{1}{R^{1-\varepsilon}}.

To prove Theorem 7.1, we start with an instance Ψ\Psi from Theorem 1.3, and first argue that Ψ\Psi can be made left regular while almost preserving soundness and completeness by deleting a small fraction of left vertices. We then use the right degree reduction technique of Moshkovitz and Raz [MR10], to obtain a bounded degree bi-regular bipartite 22-CSP.

Fix φ,ε>0\varphi,\varepsilon>0, and let Ψ\Psi to be the 22-Prover-11-Round game constructed for Theorem 1.3. Recall that this requires us to choose some large enough \ell relative to φ1,ε1\varphi^{-1},\varepsilon^{-1}, some large enough qq relative to \ell, and set R=q2R=q^{2\ell}. We also set δ=ε1000\delta=\frac{\varepsilon}{1000}, 0<c0<c arbitrarily small relative to δ\delta, and k=q2(1+c)k=q^{2(1+c)\ell}. Finally, we construct our 22-Prover-11-Round game from a hard instance of Gap3Lin with the appropriate completeness and soundness, so that it is NP-hard to distinguish between,

𝗏𝖺𝗅(Ψ)1φand𝗏𝖺𝗅(Ψ)1q2(11000δ)=1R1ε.\operatorname{{\sf val}}(\Psi)\geqslant 1-\varphi\quad\text{and}\quad\operatorname{{\sf val}}(\Psi)\leqslant\frac{1}{q^{2(1-1000\delta)\ell}}=\frac{1}{R^{1-\varepsilon}}.

It is clear that our 22-Prover-11-Round game can equivalently be viewed as an instance of bipartite 2-CSP, so let us analyze the underlying graph. Let 𝒰\mathcal{U} denote the set of possible questions to the first prover. Recall that the set of left vertices is,

𝖫𝖾𝖿𝗍={LHU|U𝒰,L𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2),LHU={0}},{\sf Left}=\{L\oplus H_{U}\;|\;\forall U\in\mathcal{U},\forall L\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2\ell),L\cap H_{U}=\{0\}\},

while the set of right vertices is

𝖱𝗂𝗀𝗁𝗍={R𝖦𝗋𝖺𝗌𝗌q(𝔽qU,2(1δ))|U𝒰}.{\ \sf Right}=\{R\in{\sf Grass}_{q}(\mathbb{F}_{q}^{U},2(1-\delta)\ell)\;|\;U\in\mathcal{U}\}.

The edges and constraints of this graph are generated by a randomized process. Equivalently, there is a weight function w()w(\cdot) over edges (LHU,R)(L\oplus H_{U},R). Recall that the weighting is defined by first choosing a uniform LHU𝖫𝖾𝖿𝗍L\oplus H_{U}\in{\sf Left}, and then R𝖱𝗂𝗀𝗁𝗍R\in{\sf Right} according to the process descrbied in Section 4.1.4. For a fixed LHUL\oplus H_{U}, define

wLHU(R)=|{LHU[LHU]|RL}||[LHU]|1[22(1δ)]q.w_{L\oplus H_{U}}(R)=\frac{|\{L^{\prime}\oplus H_{U^{\prime}}\in[L\oplus H_{U}]\;|\;R\subseteq L^{\prime}\}|}{|[L\oplus H_{U}]|}\cdot\frac{1}{\begin{bmatrix}{2\ell}\\ {2(1-\delta)\ell}\end{bmatrix}_{q}}.

This is the probability of choosing the (LHU,R)(L\oplus H_{U},R) conditioned on first choosing LHUL\oplus H_{U}. Since we choose LHU𝖫𝖾𝖿𝗍L\oplus H_{U}\in{\sf Left} uniformly, it follows that

w(LHU,R)=wLHU(R)|𝖫𝖾𝖿𝗍|.w(L\oplus H_{U},R)=\frac{w_{L\oplus H_{U}}(R)}{|{\sf Left}|}.

Define the neighborhood of a vertex as,

𝗇𝖻(LHU)={R𝖱𝗂𝗀𝗁𝗍|wLHU(R)>0}.{\sf nb}(L\oplus H_{U})=\{R\in{\sf Right}\;|\;w_{L\oplus H_{U}}(R)>0\}.

We will now attempt to remove some left vertices and obtain a bipartite, left-regular 22-CSP. To this end, we call LHUL\oplus H_{U} trivial if there is an equation eUe\in U such that for every basis of x1,,x2𝔽qUix_{1},\ldots,x_{2\ell}\in\mathbb{F}_{q}^{U_{i}} of LL, the points xix_{i} restricted to the variables in ee are of the form (α,α,α)(\alpha,\alpha,\alpha) for some α𝔽q\alpha\in\mathbb{F}_{q}.

Claim 7.2.

The fraction of LHU𝖫𝖾𝖿𝗍L\oplus H_{U}\in{\sf Left} that are trivial is at most 2q(22c)2q^{-(2-2c)\ell}.

Proof.

Fix a U𝒰U\in\mathcal{U}. Note that it suffices to show that at most 2q(22c)2q^{-(2-2c)\ell} vertices of the form LHUL\oplus H_{U} are trivial, as for each U𝒰U\in\mathcal{U}, there are an equal number of vertices LHUL\oplus H_{U}.

Write U=(x1,,x3k)U=(x_{1},\ldots,x_{3k}), where the iith equation in UU contains the variables x3i2,x3i1,x3ix_{3i-2},x_{3i-1},x_{3i}. Call these three coordinates a block, so that each x𝔽q3kx\in\mathbb{F}_{q}^{3k} consists of kk blocks of consecutive coordinates. Let us bound the fraction of LL such that LHUL\oplus H_{U} is trivial. For y1,,y2𝔽q3ky_{1},\ldots,y_{2\ell}\in\mathbb{F}_{q}^{3k}, let s(y1,,y2)s(y_{1},\ldots,y_{2\ell}) be the number of blocks where y1,,y2y_{1},\ldots,y_{2\ell} are all of the form (α,α,α)(\alpha,\alpha,\alpha) for some α𝔽q\alpha\in\mathbb{F}_{q}. Then

PrL[LHU is trivial]2Pry1,,y2[s(y1,,y2)=0],\Pr_{L}[\text{$L\oplus H_{U}$ is trivial}]\leqslant 2\Pr_{y_{1},\ldots,y_{2\ell}}[s(y_{1},\ldots,y_{2\ell})=0],

where the factor of 22 accounts for the probability that either y1,,y2y_{1},\ldots,y_{2\ell} are not linearly dependent, or span(y1,,y2)HU{0}\operatorname{span}(y_{1},\ldots,y_{2\ell})\cap H_{U}\neq\{0\}. Note that the probability that a specific block is trivial is q4q^{-4\ell}, hence by linearity of expectation we get that

𝔼y1,,y2[s(y1,,y2)]=kq4=q(22c),\mathop{\mathbb{E}}_{y_{1},\ldots,y_{2\ell}}[s(y_{1},\ldots,y_{2\ell})]=kq^{-4\ell}=q^{-(2-2c)\ell},

and therefore

Prx1,,x2[s(x1,,x2)1]q(22c).\Pr_{x_{1},\ldots,x_{2\ell}}[s(x_{1},\dots,x_{2\ell})\geqslant 1]\leqslant q^{-(2-2c)\ell}.\qed

Let Ψ\Psi^{\prime} be the instance obtained from Ψ\Psi after removing all trivial LHUL\oplus H_{U} from Ψ\Psi, so that the new instance Ψ\Psi^{\prime} does not contain any trivial vertices. Let 𝖫𝖾𝖿𝗍{\sf Left}^{\prime} denote the set of left vertices in Ψ\Psi^{\prime} and let w()w^{\prime}(\cdot) denote the weight function over edges in Ψ\Psi^{\prime}, which is given by choosing LHU𝖫𝖾𝖿𝗍L\oplus H_{U}\in{\sf Left}^{\prime} uniformly, and then choosing R𝗇𝖻(LHU)R\in{\sf nb}(L\oplus H_{U}) with probability proportional to wLHU(R)w_{L\oplus H_{U}}(R). It follows that,

w(LHU,R)=wLHU(R)|𝖫𝖾𝖿𝗍|.w^{\prime}(L\oplus H_{U},R)=\frac{w_{L\oplus H_{U}}(R)}{|{\sf Left}^{\prime}|}.

Let EE^{\prime} be a set of edges in Ψ\Psi and for notational purposes let us write w(LHU,R)=0w^{\prime}(L\oplus H_{U},R)=0 if LHUL\oplus H_{U} is trivial and not in Ψ\Psi^{\prime}. We have

w(E)4q2+2cw(E)w(E)12q2+2c.w(E)-4q^{-2+2c\ell}\leqslant w^{\prime}(E^{\prime})\leqslant\frac{w(E^{\prime})}{1-2q^{-2+2c\ell}}. (14)

The upper bound is clear from Claim 7.2. For the lower bound, we have w(E)w(E)2q2+2c12q2+2cw(E)4q2+2cw(E^{\prime})\geqslant\frac{w(E^{\prime})-2q^{-2+2c\ell}}{1-2q^{-2+2c\ell}}\geqslant w(E)-4q^{-2+2c\ell}.

It follows that Ψ\Psi^{\prime} has completeness at least 1ψ4q(22c)1-\psi-4q^{-(2-2c)\ell} and soundness at most 2q2(11000δ)\frac{2}{q^{-2(1-1000\delta)\ell}}. We now bound the size of the neighborhoods in Ψ\Psi^{\prime}.

Claim 7.3.

For each LHUL\oplus H_{U}, we have

|𝗇𝖻(LHU)|10kq6k.|{\sf nb}(L\oplus H_{U})|\leqslant 10^{k}q^{6k\ell}.
Proof.

Let U=(x1,,x3k)=(e1,,ek)U=(x_{1},\ldots,x_{3k})=(e_{1},\ldots,e_{k}) and suppose equation eie_{i} contains variables (x3i2,x3i1,x3i)(x_{3i-2},x_{3i-1},x_{3i}). Since LL is not trivial, for each ii, there must be a point vL𝔽qUv\in L\subseteq\mathbb{F}_{q}^{U} such that the values of vv restricted to the coordinates of variables (x3i2,x3i1,x3i)(x_{3i-2},x_{3i-1},x_{3i}) are not all equal. Without loss of generality, say that it is x3ix_{3i} for each 1ik1\leqslant i\leqslant k. It follows that in order to have

LHUHU=LHUHU,L\oplus H_{U}\oplus H_{U^{\prime}}=L^{\prime}\oplus H_{U}\oplus H_{U^{\prime}},

UU^{\prime} must contain an equation with the variable x3ix_{3i} for each 1ik1\leqslant i\leqslant k. Let EiE_{i} denote this set of equations for each ii. By the regularity assumptions on our 3Lin instance, |Ei|10|E_{i}|\leqslant 10 and EiEj=E_{i}\cap E_{j}=\emptyset for iji\neq j. It follows that UU^{\prime} must contain exactly one equation from each EiE_{i}, and that these form all kk equations of UU^{\prime}, so there are at most 10k10^{k} possible UU^{\prime} for which there can exist LUL^{\prime}\subseteq U^{\prime}, such that LHU[LHU]L^{\prime}\oplus H_{U^{\prime}}\in[L\oplus H_{U}]. The lemma follows from the observation that |𝖦𝗋𝖺𝗌𝗌q(3k,2(1δ))|q6k|{\sf Grass}_{q}(3k,2(1-\delta)\ell)|\leqslant q^{6k\ell}. ∎

Performing the same procedure as in [KR03, Lemma 3.4], we can turn Ψ′′\Psi^{\prime\prime} into a bipartite, left-regular 22-CSP instance without losing too much in completeness or soundness. Let Q=10kq6kQ=10^{k}q^{6k\ell} be the upper bound on neighborhood sizes in Claim 7.3.

Claim 7.4.

For any CC\in\mathbb{N}, there is a polynomial time algorithm that takes Ψ\Psi^{\prime} as input and outputs a bipartite 22-CSP Ψ′′\Psi^{\prime\prime} that is left regular with degree CQC\cdot Q, and has

  • Completeness 1φ4q(22c)1C1-\varphi-4q^{-(2-2c)\ell}-\frac{1}{C}.

  • Soundness 2q2(11000)δ+1C\frac{2}{q^{-2(1-1000)\delta\ell}}+\frac{1}{C}.

Proof.

We define Ψ′′\Psi^{\prime\prime} as follows. For each vertex LHU𝖫𝖾𝖿𝗍L\oplus H_{U}\in{\sf Left}^{\prime}, do the following. Let R1,,RmR_{1},\ldots,R_{m} be the vertices in 𝗇𝖻(LHU){\sf nb}(L\oplus H_{U}). For each 2im2\leqslant i\leqslant m, add wLHU(Ri)CQ\lfloor w_{L\oplus H_{U}}(R_{i})C\cdot Q\rfloor edges from LHUL\oplus H_{U} to RiR_{i}. Also add CQi=2mwLHU(Ri)CQC\cdot Q-\sum_{i=2}^{m}\lfloor w_{L\oplus H_{U}}(R_{i})C\cdot Q\rfloor edges from LHUL\oplus H_{U} to R1R_{1}. It is clear that Ψ′′\Psi^{\prime\prime} is left regular with degree CQC\cdot Q, and that for each RiR_{i}, there are at most wLHU(Ri)CQw_{L\oplus H_{U}}(R_{i})C\cdot Q edges between LHUL\oplus H_{U} and RiR_{i} for 2im2\leqslant i\leqslant m, and at most (wLHU(Ri)+1C)CQ\left(w_{L\oplus H_{U}}(R_{i})+\frac{1}{C}\right)C\cdot Q edges between LHUL\oplus H_{U} and R1R_{1}.

For the completeness and soundness, consider a left vertex LHUL\oplus H_{U}. Then it is clear by the above that if a labelling satisfies 1c1-c fraction of constraints involving LHUL\oplus H_{U} in Ψ\Psi^{\prime}, then in Ψ′′\Psi^{\prime\prime} the same labelling satisfies at least (1c1C)CQ\left(1-c-\frac{1}{C}\right)C\cdot Q of the edges incident to LHUL\oplus H_{U}. Similarly, if a labelling satisfies at most ss fraction of constraints involving LHUL\oplus H_{U} in Ψ\Psi^{\prime}, then it satisfies at most (1c+1C)CQ\left(1-c+\frac{1}{C}\right)C\cdot Q edges involving LHUL\oplus H_{U}. ∎

Applying Claim 7.4 with C=q10C=q^{10\ell}, we obtain a bipartite 22-CSP, Ψ′′\Psi^{\prime\prime}, that is left regular with degree q10Qq^{10\ell}Q, that still has nearly the same completeness and soundness as our original instance Ψ\Psi. We will now create a 22-regular bipartite CSP from Ψ′′\Psi^{\prime\prime}, by using the degree reduction technique of Moshkovitz and Raz [MR10].

Lemma 7.5.

[MR10] For any parameter dd, there is a polynomial time algorithm that takes a bipartite, left regular 22-CSP with left degree, d𝗅𝖾𝖿𝗍d_{{\sf left}}, completeness 1φ1-\varphi^{\prime} and soundness ss, and outputs a bipartite, (d𝗅𝖾𝖿𝗍,dd𝗅𝖾𝖿𝗍)(d_{{\sf left}},dd_{{\sf left}})-regular 22-CSP completeness 1φ1-\varphi^{\prime} and soundness s+O(d1/2)s+O\left(d^{-1/2}\right).

We are now ready to complete the proof of Theorem 7.1.

Proof of Theorem 7.1.

Applying, Lemma 7.5 to Ψ′′\Psi^{\prime\prime} with d=q10d=q^{10\ell}, we obtain a bipartite (q10Q,q20Q)(q^{10\ell}Q,q^{20\ell}Q)-regular 22-CSP, Ψ0\Psi_{0}, with completeness at least 1φ5q(22c)1-\varphi-5q^{-(2-2c)\ell} and soundness at most 3q2(11000δ)\frac{3}{q^{-2(1-1000\delta)\ell}}. By setting qq and \ell large enough relative to φ\varphi, Ψ′′\Psi^{\prime\prime} has completeness 11.1φ1-1.1\varphi. The soundness is at most, 1R10.9ε\frac{1}{R^{1-0.9\varepsilon}}. As the original φ\varphi and ε\varepsilon can be arbitrarily small positive constants, Theorem 7.1 follows. ∎

7.3.2 Sparsification

In [LM23], Lee and Manurangsi show how to conclude Theorem 1.5 from Theorem 7.1 via a sparsification procedure. We summarize the steps here. Fix the η>0\eta>0 for Theorem 1.5. Set φ=ε=0.01η\varphi=\varepsilon=0.01\eta in Theorem 7.1 and let Ψ\Psi be the resulting hard bipartite (d1,d2)(d_{1},d_{2})-regular 2-CSP, and it is NP-hard to distinguish between.

𝗏𝖺𝗅(Ψ)1φand𝗏𝖺𝗅(Ψ)1R1ε.\operatorname{{\sf val}}(\Psi)\geqslant 1-\varphi\quad\text{and}\quad\operatorname{{\sf val}}(\Psi)\leqslant\frac{1}{R^{1-\varepsilon}}.

Now, observe that the degrees of Ψ\Psi can be multiplied by arbitrary constants by copying vertices.

Lemma 7.6.

[LM23, Lemma 10] For any integers d1,d2,c1,c2d_{1},d_{2},c_{1},c_{2}, there is a polynomial time reduction from a bipartite (d1,d2)(d_{1},d_{2})-biregular CSP, Ψ\Psi, to a bipartite (c2d1d2,c1d1d2)(c_{2}d_{1}d_{2},c_{1}d_{1}d_{2})-biregular CSP Ψ\Psi^{\prime}, such that 𝗏𝖺𝗅(Ψ)=𝗏𝖺𝗅(Ψ)\operatorname{{\sf val}}(\Psi)=\operatorname{{\sf val}}(\Psi^{\prime}), and such that the left and right alphabet sizes are preserved.

It is then shown in [LM23] that one can perform a subsampling procedure to Ψ\Psi^{\prime}, that significantly lowers the degree, while not increasing the soundness or alphabet size too much.

Theorem 7.7.

[LM23, Theorem 11] For any 0<ν1<ν210<\nu_{1}<\nu_{2}\leqslant 1, such that any positive integer CC, and any sufficiently large positive integers dA,dBd0(φ,ν)d_{A},d_{B}\geqslant d_{0}(\varphi,\nu), and RR0(δ,ν,dA,dB)R\geqslant R_{0}(\delta,\nu,d_{A},d_{B}), the following holds: there is a randomized polynomial-time reduction from a bipartite (dAC,dBC)(d_{A}C,d_{B}C)-biregular 22-CSP, Ψ\Psi^{\prime}, with alphabet size at most RR, (dA,dB)(d_{A},d_{B})-bounded degree 22-CSP, Ψ′′\Psi^{\prime\prime}, such that, with probability at least 2/32/3,

  • Completeness: 𝗏𝖺𝗅(Ψ′′)𝗏𝖺𝗅(Ψ)ν1\operatorname{{\sf val}}(\Psi^{\prime\prime})\geqslant\operatorname{{\sf val}}(\Psi^{\prime})-\nu_{1},

  • Soundness: If 𝗏𝖺𝗅(Ψ′′)1Rν2\operatorname{{\sf val}}(\Psi^{\prime\prime})\leqslant\frac{1}{R^{\nu_{2}}}, then 𝗏𝖺𝗅(Ψ′′)1ν2ν1(1dA+1dB)\operatorname{{\sf val}}(\Psi^{\prime\prime})\leqslant\frac{1}{\nu_{2}-\nu_{1}}\left(\frac{1}{d_{A}}+\frac{1}{d_{B}}\right)

Putting everything together, we can prove Theorem 1.5.

Proof of Theorem 1.5.

Recall the values η\eta and dd from Theorem 1.5. Start with an instance Ψ\Psi of 22-CSP from Theorem 7.1 with φ=ε=0.01η\varphi=\varepsilon=0.01\eta and sufficiently large alphabet size RR. Then Ψ\Psi is (d1,d2)(d_{1},d_{2})-biregular, with sufficiently large alphabet size RR relative to φ1,ε1,\varphi^{-1},\varepsilon^{-1}, and dd. For such a Ψ\Psi, it is NP-hard to distinguish whether 𝗏𝖺𝗅(Ψ)=10.01η\operatorname{{\sf val}}(\Psi)=1-0.01\eta, or 𝗏𝖺𝗅(Ψ)1R10.01η\operatorname{{\sf val}}(\Psi)\leqslant\frac{1}{R^{1-0.01\eta}}.

Applying Lemma 7.6 with c1=c2=dc_{1}=c_{2}=d yields, in polynomial time, a (dd1d2,dd1d2)(dd_{1}d_{2},dd_{1}d_{2})-biregular 22-CSP, Ψ\Psi^{\prime}, with alphabet size RR and satisfying 𝗏𝖺𝗅(Ψ)=𝗏𝖺𝗅(Ψ)\operatorname{{\sf val}}(\Psi^{\prime})=\operatorname{{\sf val}}(\Psi). Next, applying Theorem 7.7, with dA=d,dB=d,C=d1d2,ν1=0.01η,ν2=1εd_{A}=d,d_{B}=d,C=d_{1}d_{2},\nu_{1}=0.01\eta,\nu_{2}=1-\varepsilon to obtain, in randomized polynomial time, a 22-CSP, Ψ′′\Psi^{\prime\prime}, with degree at most dd such that:

  • If 𝗏𝖺𝗅(Ψ)1φ\operatorname{{\sf val}}(\Psi)\geqslant 1-\varphi, then 𝗏𝖺𝗅(Ψ′′)1φν1=10.02η\operatorname{{\sf val}}(\Psi^{\prime\prime})\geqslant 1-\varphi-\nu_{1}=1-0.02\eta.

  • If 𝗏𝖺𝗅(Ψ)1R1ε\operatorname{{\sf val}}(\Psi)\leqslant\frac{1}{R^{1-\varepsilon}}, then 𝗏𝖺𝗅(Ψ′′)11εν1(1d+1d)=110.02η2d\operatorname{{\sf val}}(\Psi^{\prime\prime})\leqslant\frac{1}{1-\varepsilon-\nu_{1}}\left(\frac{1}{d}+\frac{1}{d}\right)=\frac{1}{1-0.02\eta}\cdot\frac{2}{d}.

Finally note that,

10.02ε110.02ε2dd(12ε).\frac{1-0.02\varepsilon}{\frac{1}{1-0.02\varepsilon}\cdot\frac{2}{d}}\geqslant d\left(\frac{1}{2}-\varepsilon\right).

Thus, by Theorem 7.1 and the randomized polynomial time reduction above, it follows that unless 𝖭𝖯=𝖡𝖯𝖯{\sf NP}={\sf BPP}, there is no polynomial time d(12η)d\left(\frac{1}{2}-\eta\right) approximation algorithm for 2-CSP with degree at most dd. ∎

7.4 Proof of Theorem 1.6

Combining our 22-Prover-11-Round Game in Theorem 1.3 with [Lae14, Lemma 4] and [Man19, Theorem 1], we obtain improved hardness of approximation results for Rooted kk-connectivity on undirected graphs, the vertex-connectivity survivable network design problem, and the vertex-connectivity kk-route cut problem on undirected graphs. The reduction is exactly the same as the reduction therein and we therefore omit the details.

8 Bounding the Number of Successful Zoom-outs of a Fixed Codimension

The goal of this section is to prove Lemma 5.19. Let 𝔽qn=𝔽qV\mathbb{F}_{q}^{n}=\mathbb{F}_{q}^{V} be the space that we are working in and suppose TT is a table assigning linear functions to 𝖦𝗋𝖺𝗌𝗌q(n,2){\sf Grass}_{q}(n,2\ell). We assume n2100qn\geqslant 2^{100}q^{\ell}. Let us review the set up of Lemma 5.19. Recall that we set

ξ=δ5,δ2=ξ/100,t=22+10/δ2.\xi=\delta^{5},\quad\delta_{2}=\xi/100,\quad t=2^{2+10/\delta_{2}}.

Let 𝒮={W1,,WN}\mathcal{S}=\{W_{1},\ldots,W_{N}\} be a set of codimension rr-subspaces in 𝔽qn\mathbb{F}_{q}^{n} of size

Nq100(t1)!r2ξ1.N\geqslant q^{100(t-1)!r^{2}\ell\xi^{-1}}.

For each WiW_{i}, let fi:Wi𝔽qf_{i}:W_{i}\xrightarrow[]{}\mathbb{F}_{q} be a linear function such that fi|L=T[L]f_{i}|_{L}=T[L] for at least CC-fraction of the 22\ell-subspaces LWiL\in W_{i}, where Cq2(1ξ)C\geqslant q^{-2(1-\xi)\ell}, and ξ>0\xi>0.

8.1 Step 1: Reducing to a Generic Set of Subspaces

Applying Lemma 5.8, with parameter tt as defined, we get that there exists a subspace V𝔽qnV^{\prime}\subseteq\mathbb{F}_{q}^{n} and a set of

m1N1(r+1)(t1)!qrq75rξ1\ m_{1}\geqslant\frac{N^{\frac{1}{(r+1)\cdot(t-1)!}}}{q^{r}}\geqslant q^{75r\ell\xi^{-1}}

subspaces 𝒲={W1,,Wm1}𝒮\mathcal{W}=\{W_{1},\ldots,W_{m_{1}}\}\subseteq\mathcal{S}, such that

  • Each Wi𝒲W_{i}\in\mathcal{W} is contained in VV^{\prime} and has co-dimension ss with respect to VV^{\prime}, where srs\leqslant r.

  • 𝒲\mathcal{W} is tt-generic with respect to VV^{\prime}.

We remark that this subspace VV^{\prime} will ultimately be the one used for Lemma 5.19. The remainder of the proof is devoted to finding the linear function h:V𝔽qh^{\prime}:V^{\prime}\xrightarrow[]{}\mathbb{F}_{q}, and the set 𝒲\mathcal{W}^{\prime}, which will be a subset of 𝒲\mathcal{W} above.

8.2 Step 2: Finding Local Agreement

For a subspace XX and linear assignment to XX, σ𝔽qX\sigma\in\mathbb{F}_{q}^{X}, let

X={L|XL}andX,σ={LX|T[L]|X=σ}.\mathcal{L}_{X}=\{L\in\mathcal{L}\;|\;X\subseteq L\}\quad\text{and}\quad\mathcal{L}_{X,\sigma}=\{L\in\mathcal{L}_{X}\;|\;T[L]|_{X}=\sigma\}.

Likewise, define

𝒲X={Wi𝒲|XWi}and𝒲X,σ={Wi𝒲X|fi|X=σ}.\mathcal{W}_{X}=\{W_{i}\in\mathcal{W}\;|\;X\subseteq W_{i}\}\quad\text{and}\quad\mathcal{W}_{X,\sigma}=\{W_{i}\in\mathcal{W}_{X}\;|\;f_{i}|_{X}=\sigma\}.

The first step of our proof is to find sets 𝒲X,σ\mathcal{W}_{X,\sigma} and X,σ\mathcal{L}_{X,\sigma} that have strong agreement between them, in the sense of the following lemma. The approach of this first step is similar to that of [IKW12, BDN17, MZ23]. Fix γ>0\gamma>0 to be a small constant, say γ=106\gamma=10^{-6}.

Lemma 8.1.

There exists a 2(1ξ2)2\left(1-\frac{\xi}{2}\right)\ell-dimensional subspace XX, a linear assignment, σ\sigma, to XX, and sets X,σ\mathcal{L}_{X,\sigma} and 𝒲X,σ\mathcal{W}_{X,\sigma} such that the following hold:

  • μX(X,σ)C6\mu_{X}(\mathcal{L}_{X,\sigma})\geqslant\frac{C}{6}.

  • |𝒲X,σ|m1q10r|\mathcal{W}_{X,\sigma}|\geqslant\frac{m_{1}}{q^{10r\ell}}.

  • Choosing LX,σL\in\mathcal{L}_{X,\sigma} uniformly, and Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} uniformly such that WiLW_{i}\supseteq L, we have

    PrLWi[fi|LT[L]]5γ.\Pr_{L\subseteq W_{i}}[f_{i}|_{L}\neq T[L]]\leqslant 5\gamma.
Proof.

Deferred to Appendix E

As an immediate corollary, we have the following statement. The difference between Corollary 8.2 and Lemma 8.1 is that the former we require the third condition to hold every every LX,σL\in\mathcal{L}_{X,\sigma}, instead of a random LL, and we also require every LX,σL\in\mathcal{L}_{X,\sigma} to be contained in roughly the same number of Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}.

Corollary 8.2.

Taking X,σ\mathcal{L}_{X,\sigma} and 𝒲X,σ\mathcal{W}_{X,\sigma} from Lemma 8.1, there is a subset X,σX,σ\mathcal{L}^{\prime}_{X,\sigma}\subseteq\mathcal{L}_{X,\sigma} such that the following hold.

  • μX(X,σ)C12\mu_{X}(\mathcal{L}^{\prime}_{X,\sigma})\geqslant\frac{C}{12}

  • m2=|𝒲X,σ|m1q10rm_{2}=|\mathcal{W}_{X,\sigma}|\geqslant\frac{m_{1}}{q^{10r\ell}}.

  • For every LX,σL\in\mathcal{L}^{\prime}_{X,\sigma}, choosing Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} uniformly such that WiLW_{i}\supseteq L, we have

    PrWiL,Wi𝒲X,σ[fi|LT[L]]12γ.\Pr_{W_{i}\supseteq L,W_{i}\in\mathcal{W}_{X,\sigma}}[f_{i}|_{L}\neq T[L]]\leqslant 12\gamma.
  • For every LX,σL\in\mathcal{L}^{\prime}_{X,\sigma},

    0.95|𝒲X,σ|qξsN𝒲X,σ(L)1.05|𝒲X,σ|qξs,0.95\cdot|\mathcal{W}_{X,\sigma}|\cdot q^{-\xi\ell\cdot s}\leqslant N_{\mathcal{W}_{X,\sigma}}(L)\leqslant 1.05\cdot|\mathcal{W}_{X,\sigma}|\cdot q^{-\xi\ell\cdot s},

where N𝒲X,σ(L)=|{WiL|Wi𝒲X,σ}|N_{\mathcal{W}_{X,\sigma}}(L)=|\{W_{i}\supseteq L\;|\;W_{i}\in\mathcal{W}_{X,\sigma}\}|.

Proof.

Take X,σ,X,σ,X,\sigma,\mathcal{L}_{X,\sigma}, and 𝒲X,σ\mathcal{W}_{X,\sigma} as guaranteed by Lemma 8.1, so that μX(X,σ)C6\mu_{X}(\mathcal{L}_{X,\sigma})\geqslant\frac{C}{6}. We will keep the same X,σ,X,\sigma,, but we remove some LL’s from X,σ\mathcal{L}_{X,\sigma} to make the third and fourth items hold.

By Markov’s inequality, at most 512\frac{5}{12}-fraction of LX,σL\in\mathcal{L}_{X,\sigma} violate the third item. By Lemma 5.16 applied to 𝒲X,σ\mathcal{W}_{X,\sigma} with parameters Q=XQ=X, j=2j=2\ell, a=2(1ξ2)a=2\left(1-\frac{\xi}{2}\right)\ell, r=sr=s, and c=0.05c=0.05, we have that

PrL𝖹𝗈𝗈𝗆[Q,V][|NX,σ(L)qξsm2|0.1qξm2]400qξsm2C100,\Pr_{L\in{\sf Zoom}[Q,V]}\left[\left|N_{X,\sigma}(L)-q^{-\xi\ell\cdot s}m_{2}\right|\geqslant 0.1q^{-\xi\ell}m_{2}\right]\leqslant\frac{400q^{-\xi\ell\cdot s}}{m_{2}}\leqslant\frac{C}{100},

where NX,σ(L)=|{WiL|Wi𝒲X,σ}|N_{X,\sigma}(L)=|\{W_{i}\supseteq L\;|\;W_{i}\in\mathcal{W}_{X,\sigma}\}|. It follows that after removing the LX,σL\in\mathcal{L}_{X,\sigma} that do not satisfy the third or fourth condition, we arrive at the desired X,σ\mathcal{L}^{\prime}_{X,\sigma}, which still has measure at least

712C6C100C12.\frac{7}{12}\cdot\frac{C}{6}-\frac{C}{100}\geqslant\frac{C}{12}.\qed

We fix X,σX,\sigma as well as WX,σW_{X,\sigma} and X,σ\mathcal{L}^{\prime}_{X,\sigma} as in Corollary 8.2 throughout the rest of the argument.

8.3 Step 3: A Global Set with Local Agreement

The next step is to further refine the set X,σ\mathcal{L}^{\prime}_{X,\sigma} so that the remaining subspaces “evenly cover” a subspace of VVV^{\star}\subseteq V^{\prime} with codimension dim(X)+Oδ2(1)\dim(X)+O_{\delta_{2}}(1). To do this, we will reduce to the case where X,σ\mathcal{L}^{\prime}_{X,\sigma} is global within some zoom-in AA and zoom-out BB such that XABX\subseteq A\subseteq B. This is done via the following argument, which we outline below:

  1. 1.

    While X,σ\mathcal{L}^{\prime}_{X,\sigma} is not global, there must be some zoom-in or zoom-out on which it is dense, so consider the restriction to this zoom-in or zoom-out.

  2. 2.

    This increases the measure of X,σ\mathcal{L}^{\prime}_{X,\sigma}, and we may repeat until we have a global set (within some zoom-in, zoom-out combination).

  3. 3.

    By choosing the globalness parameter suitably, we are able to perform the above process in relatively few times until the restriction of X,σ\mathcal{L}_{X,\sigma}^{\prime} that we arrive at is global.

  4. 4.

    As a result, when restricting to the zoom-in, zoom-out combination, the resulting set of subspaces evenly covers a space that is still relatively large in the sense that it contains qO(ξ1)q^{-O(\xi^{-1})}-fraction of V0V_{0}.

For a zoom-in AA and zoom-out BB such that XABVX\subseteq A\subseteq B\subseteq V^{\prime}, write V=AV0V^{\prime}=A\oplus V_{0} and B=AVB=A\oplus V^{\star}, where VV0V^{\star}\subseteq V_{0}. Also define

𝒲[A,B]={Wi|Wi𝒲X,σ s.t AWi=WiB}.\mathcal{W}^{\star}_{[A,B]}=\{W^{\star}_{i}\;|\;\exists W_{i}\in\mathcal{W}_{X,\sigma}\text{ s.t }A\oplus W^{\star}_{i}=W_{i}\cap B\}.

It is clear that each Wi𝒲[A,B]W^{\star}_{i}\in\mathcal{W}^{\star}_{[A,B]} is contained inside of some Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}, so for each WiW^{\star}_{i} we may define the restriction of fif_{i} to WiW^{\star}_{i} as fi=fi|Wif^{\star}_{i}=f_{i}|_{W^{\star}_{i}}.

Lemma 8.3.

There is a zoom-in AA and a zoom-out BB such that the following holds with the notation above. There exists a collection of subspaces 𝒲={W1,,Wm3}𝒲[A,B]\mathcal{W}^{\star}=\{W^{\star}_{1},\ldots,W^{\star}_{m_{3}}\}\subseteq\mathcal{W}^{\star}_{[A,B]} of codimension ss with respect to VV^{\star} such that:

  1. 1.

    For some ξ3\ell^{\prime}\geqslant\frac{\xi}{3}\ell there exists 𝖦𝗋𝖺𝗌𝗌q(V,)\mathcal{L}^{\star}\subseteq{\sf Grass}_{q}(V^{\star},\ell^{\prime}) such that μ()=ηC12\mu(\mathcal{L}^{\star})=\eta\geqslant\frac{C}{12}.

  2. 2.

    The set \mathcal{L}^{\star} is (1,qδ2η)(1,q^{\delta_{2}\ell}\eta)-pseudo-random.

  3. 3.

    Each WiW^{\star}_{i} has codimension srs\leqslant r inside of VV^{\star} and 𝒲\mathcal{W}^{\star} is 44-generic, with respect to VV^{\star}.

  4. 4.

    m3q10s/δ22m2m_{3}\geqslant\;\frac{q^{-10s/\delta_{2}}}{2}\cdot m_{2}.

  5. 5.

    For every LL\in\mathcal{L}^{\star}, choosing Wi𝒲W^{\star}_{i}\in\mathcal{W}^{\star} uniformly such that WiLW^{\star}_{i}\supseteq L, we have

    PrWiL,Wi𝒲[fi|LT[L]]14γ.\Pr_{W^{\star}_{i}\supseteq L,W^{\star}_{i}\in\mathcal{W}^{\star}}[f_{i}|_{L}\neq T[L]]\leqslant 14\gamma.
  6. 6.

    For every LL\in\mathcal{L}^{\star},

    0.8m3qsN𝒲(L)1.2m3qs,0.8\cdot m_{3}\cdot q^{-s\cdot\ell^{\prime}}\leqslant N_{\mathcal{W}^{\star}}(L)\leqslant 1.2\cdot m_{3}\cdot q^{-s\cdot\ell^{\prime}},

    where N𝒲(L)=|{Wi𝒲|WiL}|N_{\mathcal{W}^{\star}}(L)=|\{W^{\star}_{i}\in\mathcal{W}^{\star}\;|\;W^{\star}_{i}\supseteq L\}|

Here the table TT is assigns linear functions to L𝖦𝗋𝖺𝗌𝗌q(V,)L\in{\sf Grass}_{q}(V^{\star},\ell), and is essentially the original table, i.e

T[L]T[AL]|L.T[L]\equiv T[A\oplus L]|_{L}.
Proof.

Deferred to Appendix E.2. ∎

Finally, as a consequence of pseudo-randomness, we may apply Lemma 2.4, to get that \mathcal{L}^{\star} evenly covers VV^{\star}.

Lemma 8.4.

Setting Z={zV||μz()η|η10}Z=\{z\in V^{\star}\;|\;|\mu_{z}(\mathcal{L}^{\star})-\eta|\leqslant\frac{\eta}{10}\}, we have that,

|Z||V|1q2.\frac{|Z|}{|V^{\star}|}\geqslant 1-q^{\frac{\ell^{\prime}}{2}}.
Proof.

This is immediate by the pseudo-randomness of \mathcal{L}^{\star} and Lemma 2.4. ∎

8.4 Step 4: Local to Global Agreement

Lemma 8.5.

We have

PrWi,Wj𝒲zWiWjZ[fi(z)fj(z)]500γ,\Pr_{\begin{subarray}{c}W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star}\\ z\in W^{\star}_{i}\cap W^{\star}_{j}\cap Z\end{subarray}}[f^{\star}_{i}(z)\neq f^{\star}_{j}(z)]\leqslant 500\gamma,

and for every Wi,Wj𝒲W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star},

|WiWjZ|0.81|WiWj|.|W^{\star}_{i}\cap W^{\star}_{j}\cap Z|\geqslant 0.81\cdot|W^{\star}_{i}\cap W^{\star}_{j}|.
Proof.

Deferred to Section E.3. ∎

Using Lemma 8.5, we conclude the proof of Lemma 5.19 by using ideas from the Raz-Safra analysis of the Plane versus Plane test [RS97]. Define the following graph, GG, with vertex set 𝒲\mathcal{W}^{\star} and an edge between Wi,WjW^{\star}_{i},W^{\star}_{j} if and only if fi|WiWj=fj|WiWjf^{\star}_{i}|_{W^{\star}_{i}\cap W^{\star}_{j}}=f^{\star}_{j}|_{W^{\star}_{i}\cap W^{\star}_{j}}. We claim that this graph contains a large clique. To do so, we show that the graph is nearly transitive. For a graph H=(V,E)H=(V,E), define

β(H)=max(u,w)EPrv[(v,u),(v,w)E].\beta(H)=\max_{(u,w)\notin E}\Pr_{v}[(v,u),(v,w)\in E].

A graph HH is transitive β(H)=0\beta(H)=0. It is easy to see that transitive graphs are (edge) disjoint unions of cliques. The following lemma, proved in [RS97], asserts that if HH is relatively dense and β(H)\beta(H) is small, then one could remove only a small fraction of the edges and get a fully transitive graph.

Lemma 8.6.

[RS97, Lemma 2] Any graph H=(V,E)H=(V,E) can be made transitive by deleting at most 3β(H)|V|23\sqrt{\beta(H)}|V|^{2} edges.

To use lemma 8.6 we first show that the graph GG we defined is highly transitive.

Claim 8.7.

We have β(G)1m3\beta(G)\leqslant\frac{1}{m_{3}}.

Proof.

Fix a Wi,WjW^{\star}_{i},W^{\star}_{j} that are not adjacent. We claim that they can have at most 11 common neighbor. Suppose for the sake of contradiction that Wa,WbW^{\star}_{a},W^{\star}_{b} are distinct common neighbors. Then,

fi|WiWjWa=fj|WiWjWa,f^{\star}_{i}|_{W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{a}}=f^{\star}_{j}|_{W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{a}},

and

fi|WiWjWb=fj|WiWjWb.f^{\star}_{i}|_{W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{b}}=f^{\star}_{j}|_{W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{b}}.

It follows that fif^{\star}_{i} and fjf^{\star}_{j} agree on WiWjWaWiWjWbW^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{a}\oplus W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{b}. However, since 𝒲\mathcal{W}^{\star} is 44-generic, we have

codim(WiWjWaWiWjWb)3s+3s4s=2s\operatorname{codim}(W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{a}\oplus W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{b})\leqslant 3s+3s-4s=2s

and

WiWjWaWiWjWbWiWj,W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{a}\oplus W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{b}\subseteq W^{\star}_{i}\cap W^{\star}_{j},

so it must be the case that WiWjWaWiWjWb=WiWjW^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{a}\oplus W^{\star}_{i}\cap W^{\star}_{j}\cap W^{\star}_{b}=W^{\star}_{i}\cap W^{\star}_{j}. This contradicts the assumption that WiW^{\star}_{i} and WjW^{\star}_{j} are not adjacent. Thus, any two non-adjacent vertices can have at most 11 common neighbor, and the result follows. ∎

Claim 8.8.

The graph GG contains a clique of size of size m32\frac{m_{3}}{2}

Proof.

Applying Markov’s inequality and a union bound to Lemma 8.5, we have that with probability at least 9/109/10 over WiW^{\star}_{i} and WjW^{\star}_{j}, we have both PrzWiWjZ[fi(z)fj(z)]10001γ\Pr_{z\in W^{\star}_{i}\cap W^{\star}_{j}\cap Z}[f^{\star}_{i}(z)\neq f^{\star}_{j}(z)]\leqslant 10001\gamma and |WiWjZ|0.81|WiWj||W^{\star}_{i}\cap W^{\star}_{j}\cap Z|\geqslant 0.81\cdot|W^{\star}_{i}\cap W^{\star}_{j}|. In this case, fif^{\star}_{i} and fjf^{\star}_{j} agree on at least (110001γ)(1-10001\gamma)-fraction of the points in WiWjZW^{\star}_{i}\cap W^{\star}_{j}\cap Z, which is in turn at least (110001γ)0.8>1/q(1-10001\gamma)\cdot 0.8>1/q-fraction of the points in WiWjW^{\star}_{i}\cap W^{\star}_{j}. As fif^{\star}_{i} and fjf^{\star}_{j} are linear functions, the Schwartz-Zippel lemma implies that such Wi,WjW^{\star}_{i},W^{\star}_{j} are adjacent in GG and that GG has at least 81m32/10081m_{3}^{2}/100 edges.

By Claim 8.7 and Lemma 8.6, we can delete 2m33/22m_{3}^{3/2} edges to make GG a union of cliques. Doing so yields a graph on m3m_{3} vertices with at least m32/2m_{3}^{2}/2-edges that is a union of cliques. Let C1,,CNC_{1},\ldots,C_{N} be the cliques, with C1C_{1} being the largest one. We have,

|C1|m|C1|i=1N|Ci|i=1N|Ci|2m322.|C_{1}|\cdot m\geqslant|C_{1}|\cdot\sum_{i=1}^{N}|C_{i}|\geqslant\sum_{i=1}^{N}|C_{i}|^{2}\geqslant\frac{m_{3}^{2}}{2}.

It follows that |C1|m32|C_{1}|\geqslant\frac{m_{3}}{2}, and that GG contains a clique of size at least m32\frac{m_{3}}{2}. ∎

Let 𝒞\mathcal{C} be the clique guaranteed by Claim 8.8 and write 𝒞={W1,,Wm32}\mathcal{C}=\{W^{\star}_{1},\ldots,W^{\star}_{\frac{m_{3}}{2}}\}. To complete the proof of Lemma 5.19, we will find a linear hh, such that for all 1im321\leqslant i\leqslant\frac{m_{3}}{2}, fi|Vh|Vf^{\star}_{i}|_{V^{\star}}\equiv h|_{V^{\star}}, and then show that this hh can be extended to XAVX\oplus A\oplus V^{\star} in a manner that is consistent with many of the original fif_{i}’s for 1im3/21\leqslant i\leqslant m_{3}/2. To this end, first define g:V𝔽qg:V^{\star}\xrightarrow[]{}\mathbb{F}_{q} as follows:

g(x)={fi(x),if Wi𝒞,xWi0,otherwise.g(x)=\begin{cases}f^{\star}_{i}(x),&\text{if }\ \exists W^{\star}_{i}\in\mathcal{C},x\in W^{\star}_{i}\\ 0,&\text{otherwise.}\end{cases} (15)

Since fi(x)=fj(x)f^{\star}_{i}(x)=f^{\star}_{j}(x) whenever xWiWjx\in W^{\star}_{i}\cap W^{\star}_{j}, it does not matter which ii is chosen if there are multiple Wi𝒞W^{\star}_{i}\in\mathcal{C} containing xx. Thus, gg is well defined and g|Wi=fig|_{W^{\star}_{i}}=f^{\star}_{i} for all 1im321\leqslant i\leqslant\frac{m_{3}}{2}.

We next show that gg is close to a linear function and that this linear function agrees with most of the functions fi|Wif^{\star}_{i}|_{W^{\star}_{i}} for Wi𝒞W^{\star}_{i}\in\mathcal{C}. To begin, we show that gg passes the standard linearity test with high probability.

Lemma 8.9.

We have,

Prz1,z2𝔽qn[g(z1+z2)=g(z1)+g(z2)]13q2sm3.\Pr_{z_{1},z_{2}\in\mathbb{F}_{q}^{n}}[g(z_{1}+z_{2})=g(z_{1})+g(z_{2})]\geqslant 1-\frac{3q^{2s}}{m_{3}}.
Proof.

Note that we have

Prz1,z2𝔽qn[g(z1+z2)=g(z1)+g(z2)]Prz1,z2𝔽qn[Wi𝒞, s.t. ,z1,z2Wi].\Pr_{z_{1},z_{2}\in\mathbb{F}_{q}^{n}}[g(z_{1}+z_{2})=g(z_{1})+g(z_{2})]\geqslant\Pr_{z_{1},z_{2}\in\mathbb{F}_{q}^{n}}[\exists W^{\star}_{i}\in\mathcal{C},\text{ s.t. },z_{1},z_{2}\in W^{\star}_{i}].

For every z1,z2𝔽qnz_{1},z_{2}\in\mathbb{F}_{q}^{n} linearly independent, we can let N(z1,z2)N(z_{1},z_{2}) denote the number of Wi𝒞W^{\star}_{i}\in\mathcal{C} containing span(z1,z2)\operatorname{span}(z_{1},z_{2}). The result them follows from Lemma 5.16 with a=0a=0, j=2j=2, r=sr=s, and c=0.99c=0.99. We have,

Prz1,z2𝔽qn[Wi𝒞, s.t. ,z1,z2Wi]\displaystyle Pr_{z_{1},z_{2}\in\mathbb{F}_{q}^{n}}[\exists W^{\star}_{i}\in\mathcal{C},\text{ s.t. },z_{1},z_{2}\in W^{\star}_{i}] Prz1,z2𝔽qn[N(z1,z2)>0|dim(span(z1,z2))=2]qqn\displaystyle\geqslant\Pr_{z_{1},z_{2}\in\mathbb{F}_{q}^{n}}[N(z_{1},z_{2})>0\;|\;\dim(\operatorname{span}(z_{1},z_{2}))=2]-\frac{q}{q^{n}}
13q3sm3,\displaystyle\geqslant 1-\frac{3q^{3s}}{m_{3}},

where qqn\frac{q}{q^{n}} is an upper bound on the probability that dim(span(z1,z2))2\dim(\operatorname{span}(z_{1},z_{2}))\neq 2. ∎

Applying the linearity testing result of Blum, Luby, and Rubinfeld [BLR93, Theorem 4.1] we get that gg is 12q2sm3\frac{12q^{2s}}{m_{3}}-close to a linear function, say h:V𝔽qh:V^{\star}\xrightarrow[]{}\mathbb{F}_{q}. We will conclude by showing that this hh is the desired function which agrees with many of the original fif_{i}’s. To this end, we first show that agrees with many of the fif^{\star}_{i}’s that we have (which are restrictions of the original fif_{i}’s), and then show that hh can be extended to VV^{\prime} in a manner that retains agreement with many of the fif_{i}’s.

Towards the first step, set S={xV|g(x)h(x)}S=\{x\in V^{\star}\;|\;g(x)\neq h(x)\}. We show that choosing Wi𝒞W^{\star}_{i}\in\mathcal{C} randomly, and then a point xWix\in W^{\star}_{i}, it is unlikely that xSx\in S. Define the measure ν\nu over nonzero points in 𝔽qn\mathbb{F}_{q}^{n} obtained by choosing Wi𝒞W_{i}\in\mathcal{C} uniformly at random and then xWix\in W_{i} nonzero uniformly at random. Let μ\mu be the uniform measure, so μ(S)3q2sm3\mu(S)\leqslant\frac{3q^{2s}}{m_{3}}. Then ν(S)\nu(S) is precisely the probability of interest and can be upper bounded using Lemma 5.14, with parameters a=0,j=1a=0,j=1, codimension ss,

ν(S)μ(S)+6qs2m37qs2m3.\nu(S)\leqslant\mu(S)+\frac{6q^{\frac{s}{2}}}{\sqrt{m_{3}}}\leqslant\frac{7q^{\frac{s}{2}}}{\sqrt{m_{3}}}. (16)
Lemma 8.10.

We have h|Wifih|_{W^{\star}_{i}}\equiv f^{\star}_{i} for at least half of the Wi𝒞W^{\star}_{i}\in\mathcal{C}.

Proof.

By Markov’s inequality and Equation (16) with probability at least 1/21/2, over Wi𝒞W^{\star}_{i}\in\mathcal{C}, we have

|WiS||Wi|14qs2m3<11q,\frac{\left|W^{\star}_{i}\cap S\right|}{\left|W^{\star}_{i}\right|}\leqslant\frac{14q^{\frac{s}{2}}}{\sqrt{m_{3}}}<1-\frac{1}{q},

and fif^{\star}_{i} and h|Wih|_{W^{\star}_{i}} agree on more than 1/q1/q of the points in WiW^{\star}_{i}. Since fif^{\star}_{i} and h|Wih|_{W^{\star}_{i}} are both linear, by the Schwartz-Zippel Lemma that h|Wifih|_{W^{\star}_{i}}\equiv f^{\star}_{i}, and the result follows. ∎

We are now ready to finish the proof of Lemma 5.19.

Proof of Lemma 5.19.

Summarizing, we now have linear functions fi:Wi𝔽qf^{\star}_{i}:W^{\star}_{i}\xrightarrow[]{}\mathbb{F}_{q} for 1im341\leqslant i\leqslant\frac{m_{3}}{4} and a linear function h:V𝔽qh:V^{\star}\xrightarrow[]{}\mathbb{F}_{q} such that h|Wi=fih|_{W^{\star}_{i}}=f^{\star}_{i}. Furthermore, for each fi,Wif^{\star}_{i},W^{\star}_{i}, there is a fi,Wif_{i},W_{i} from Lemma 5.19such that WiV=WiW_{i}\cap V^{\star}=W^{\star}_{i}, WiVW_{i}\subseteq V^{\prime}, fi|Wi=fif_{i}|_{W^{\star}_{i}}=f^{\star}_{i}, and fi|X=σf_{i}|_{X}=\sigma.

Finally, we will extend hh in a manner so that it agrees with many of these original functions fif_{i}. To this end, recall that VV^{\star} satisfies,

AV=BV.A\oplus V^{\star}=B\subseteq V^{\prime}.

and dim(A)+codim(B)dim(X)10δ2\dim(A)+\operatorname{codim}(B)-\dim(X)\leqslant\frac{10}{\delta_{2}}. Therefore, we may choose a random linear function h:V𝔽qh^{\prime}:V^{\prime}\xrightarrow[]{}\mathbb{F}_{q} conditioned on h|Vhh|_{V^{\star}}\equiv h and h|Xσh^{\prime}|_{X}\equiv\sigma. For any fif_{i}, we have that

Prh[h|Wifi]q(dim(A)+codim(B)dim(X))q10δ2.\Pr_{h^{\prime}}[h^{\prime}|_{W_{i}}\equiv f_{i}]\leqslant q^{-(\dim(A)+\operatorname{codim}(B)-\dim(X))}\leqslant q^{-\frac{10}{\delta_{2}}}.

Indeed there is a q(dim(A)dim(X))q^{-(\dim(A)-\dim(X))} probability that h|Afi|Ah|_{A}\equiv f_{i}|_{A}, as we condition on h|Xσfi|Xh^{\prime}|_{X}\equiv\sigma\equiv f_{i}|_{X}. Then, extending hh^{\prime} from BB to VV^{\prime}, there is at least a qcodim(B)q^{-\operatorname{codim}(B)} probability that hh^{\prime} is equal to fif_{i} on these extra dimensions. It follows that there is a linear h:V𝔽qh^{\prime}:V^{\prime}\xrightarrow[]{}\mathbb{F}_{q} such that h|Wifih^{\prime}|_{W_{i}}\equiv f_{i} for at least

m3q10δ24q50rξ1\frac{m_{3}q^{-\frac{10}{\delta_{2}}}}{4}\geqslant q^{50r\ell\xi^{-1}}

of the pairs fi,Wif_{i},W_{i} from Lemma 5.19. Take these WiW_{i} to be the set 𝒲\mathcal{W}^{\prime} for Lemma 5.19. As these 𝒲𝒲\mathcal{W}^{\prime}\subseteq\mathcal{W}, they are 22-generic with respect to VV^{\prime} and have codimension srs\leqslant r in VV^{\prime}. ∎

9 Acknowledgements

We thank the anonymous reviewers for their comments.

References

  • [ABH+05] Sanjeev Arora, Eli Berger, Elad Hazan, Guy Kindler, and Muli Safra. On non-approximability for quadratic programs. In 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2005), 23-25 October 2005, Pittsburgh, PA, USA, Proceedings, pages 206–215. IEEE Computer Society, 2005.
  • [ALM+92] Sanjeev Arora, Carsten Lund, Rajeev Motwani, Madhu Sudan, and Mario Szegedy. Proof verification and hardness of approximation problems. In 33rd Annual Symposium on Foundations of Computer Science, Pittsburgh, Pennsylvania, USA, 24-27 October 1992, pages 14–23, 1992.
  • [AS03] Sanjeev Arora and Madhu Sudan. Improved low-degree testing and its applications. Comb., 23(3):365–426, 2003.
  • [BDN17] Amey Bhangale, Irit Dinur, and Inbal Livni Navon. Cube vs. cube low degree test. In Christos H. Papadimitriou, editor, 8th Innovations in Theoretical Computer Science Conference, ITCS 2017, January 9-11, 2017, Berkeley, CA, USA, volume 67 of LIPIcs, pages 40:1–40:31. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2017.
  • [Ber00] Piotr Berman. A d/2 approximation for maximum weight independent set in d-claw free graphs. Nordic Journal of Computing, 7(3):178–184, 2000.
  • [BKS19] Boaz Barak, Pravesh K. Kothari, and David Steurer. Small-set expansion in shortcode graph and the 2-to-2 conjecture. In Avrim Blum, editor, 10th Innovations in Theoretical Computer Science Conference, ITCS 2019, January 10-12, 2019, San Diego, California, USA, volume 124 of LIPIcs, pages 9:1–9:12. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2019.
  • [BL15] Abhishek Bhowmick and Shachar Lovett. The list decoding radius of reed-muller codes over small fields. In Rocco A. Servedio and Ronitt Rubinfeld, editors, Proceedings of the Forty-Seventh Annual ACM on Symposium on Theory of Computing, STOC 2015, Portland, OR, USA, June 14-17, 2015, pages 277–285. ACM, 2015.
  • [BLR93] Manuel Blum, Michael Luby, and Ronitt Rubinfeld. Self-testing/correcting with applications to numerical problems. J. Comput. Syst. Sci., 47(3):549–595, 1993.
  • [CCK08] Tanmoy Chakraborty, Julia Chuzhoy, and Sanjeev Khanna. Network design for vertex connectivity. In Cynthia Dwork, editor, Proceedings of the 40th Annual ACM Symposium on Theory of Computing, Victoria, British Columbia, Canada, May 17-20, 2008, pages 167–176. ACM, 2008.
  • [CDGT22] Julia Chuzhoy, Mina Dalirrooyfard, Vadim Grinberg, and Zihan Tan. A new conjecture on hardness of low-degree 2-csp’s with implications to hardness of densest k-subgraph and other problems. CoRR, abs/2211.05906, 2022.
  • [CGM13] Marek Cygan, Fabrizio Grandoni, and Monaldo Mastrolilli. How to sell hyperedges: The hypermatching assignment problem. In Sanjeev Khanna, editor, Proceedings of the Twenty-Fourth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2013, New Orleans, Louisiana, USA, January 6-8, 2013, pages 342–351. SIAM, 2013.
  • [Cha16] Siu On Chan. Approximation resistance from pairwise-independent subgroups. Journal of the ACM (JACM), 63(3):1–32, 2016.
  • [CW04] Moses Charikar and Anthony Wirth. Maximizing quadratic programs: Extending grothendieck’s inequality. In 45th Symposium on Foundations of Computer Science (FOCS 2004), 17-19 October 2004, Rome, Italy, Proceedings, pages 54–60. IEEE Computer Society, 2004.
  • [DFRR23] Pavel Dvořák, Andreas Emil Feldmann, Ashutosh Rai, and Paweł Rzǎżewski. Parameterized inapproximability of independent set in h-free graphs. Algorithmica, 85(4):902–928, 2023.
  • [DKK+18] Irit Dinur, Subhash Khot, Guy Kindler, Dor Minzer, and Muli Safra. Towards a proof of the 2-to-1 games conjecture? In Ilias Diakonikolas, David Kempe, and Monika Henzinger, editors, Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2018, Los Angeles, CA, USA, June 25-29, 2018, pages 376–389. ACM, 2018.
  • [DKK+21] Irit Dinur, Subhash Khot, Guy Kindler, Dor Minzer, and Muli Safra. On non-optimally expanding sets in grassmann graphs. Israel Journal of Mathematics, 243(1):377–420, 2021.
  • [EKL23] David Ellis, Guy Kindler, and Noam Lifshitz. An analogue of bonami’s lemma for functions on spaces of linear maps, and 2-2 games. In Barna Saha and Rocco A. Servedio, editors, Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, Orlando, FL, USA, June 20-23, 2023, pages 656–660. ACM, 2023.
  • [EKL24] Shai Evra, Guy Kindler, and Noam Lifshitz. Polynomial bogolyubov for special linear groups via tensor rank, 2024. arXiv:2404.00641.
  • [Fei98] Uriel Feige. A threshold of ln n for approximating set cover. J. ACM, 45(4):634–652, 1998.
  • [FGL+91] Uriel Feige, Shafi Goldwasser, László Lovász, Shmuel Safra, and Mario Szegedy. Approximating clique is almost np-complete (preliminary version). In 32nd Annual Symposium on Foundations of Computer Science, San Juan, Puerto Rico, 1-4 October 1991, pages 2–12, 1991.
  • [Gop13] Parikshit Gopalan. A fourier-analytic approach to reed-muller decoding. IEEE Trans. Inf. Theory, 59(11):7747–7760, 2013.
  • [GRS00] Oded Goldreich, Ronitt Rubinfeld, and Madhu Sudan. Learning polynomials with queries: The highly noisy case. SIAM J. Discret. Math., 13(4):535–570, 2000.
  • [Hås96] Johan Håstad. Clique is hard to approximate within n1-epsilon{}^{\mbox{1-epsilon}}. In 37th Annual Symposium on Foundations of Computer Science, FOCS ’96, Burlington, Vermont, USA, 14-16 October, 1996, pages 627–636. IEEE Computer Society, 1996.
  • [Hås01] Johan Håstad. Some optimal inapproximability results. J. ACM, 48(4):798–859, 2001.
  • [IKW12] Russell Impagliazzo, Valentine Kabanets, and Avi Wigderson. New direct-product testers and 2-query PCPs. SIAM J. Comput., 41(6):1722–1768, 2012.
  • [KMS17] Subhash Khot, Dor Minzer, and Muli Safra. On independent sets, 2-to-2 games, and grassmann graphs. In Hamed Hatami, Pierre McKenzie, and Valerie King, editors, Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2017, Montreal, QC, Canada, June 19-23, 2017, pages 576–589. ACM, 2017.
  • [KMS23] Subhash Khot, Dor Minzer, and Muli Safra. Pseudorandom sets in grassmann graph have near-perfect expansion. Annals of Mathematics, 198(1):1–92, 2023.
  • [KP06] Subhash Khot and Ashok Kumar Ponnuswami. Better inapproximability results for maxclique, chromatic number and min-3lin-deletion. In Michele Bugliesi, Bart Preneel, Vladimiro Sassone, and Ingo Wegener, editors, Automata, Languages and Programming, 33rd International Colloquium, ICALP 2006, Venice, Italy, July 10-14, 2006, Proceedings, Part I, volume 4051 of Lecture Notes in Computer Science, pages 226–237. Springer, 2006.
  • [KR03] Subhash Khot and Oded Regev. Vertex cover might be hard to approximate to within 2-\varepsilon. In 18th Annual IEEE Conference on Computational Complexity (Complexity 2003), 7-10 July 2003, Aarhus, Denmark, page 379. IEEE Computer Society, 2003.
  • [KS13] Subhash Khot and Muli Safra. A two-prover one-round game with strong soundness. Theory Comput., 9:863–887, 2013.
  • [Lae14] Bundit Laekhanukit. Parameters of two-prover-one-round game and the hardness of connectivity problems. In Chandra Chekuri, editor, Proceedings of the Twenty-Fifth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2014, Portland, Oregon, USA, January 5-7, 2014, pages 1626–1643. SIAM, 2014.
  • [LM23] Euiwoong Lee and Pasin Manurangsi. Hardness of approximating bounded-degree max 2-csp and independent set on k-claw-free graphs. CoRR, abs/2309.04099, 2023.
  • [Man19] Pasin Manurangsi. A note on degree vs gap of min-rep label cover and improved inapproximability for connectivity problems. Inf. Process. Lett., 145:24–29, 2019.
  • [Meg01] Alexandre Megretski. Relaxations of quadratic programs in operator theory and system analysis. In Systems, Approximation, Singular Integral Operators, and Related Topics: International Workshop on Operator Theory and Applications, IWOTA 2000, pages 365–392. Springer, 2001.
  • [MR10] Dana Moshkovitz and Ran Raz. Two-query PCP with subconstant error. J. ACM, 57(5):29:1–29:29, 2010.
  • [MZ23] Dor Minzer and Kai Zheng. Approaching the soundness barrier: A near optimal analysis of the cube versus cube test. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 2761–2776. SIAM, 2023.
  • [NRT99] Arkadi Nemirovski, Cornelis Roos, and Tamás Terlaky. On maximization of quadratic form over intersection of ellipsoids with common center. Mathematical programming, 86(3):463–473, 1999.
  • [Nut12] Zeev Nutov. Approximating minimum-cost connectivity problems via uncrossable bifamilies. ACM Transactions on Algorithms (TALG), 9(1):1–16, 2012.
  • [Rao11] Anup Rao. Parallel repetition in projection games and a concentration bound. SIAM J. Comput., 40(6):1871–1891, 2011.
  • [Raz98] Ran Raz. A parallel repetition theorem. SIAM J. Comput., 27(3):763–803, 1998.
  • [RS97] Ran Raz and Shmuel Safra. A sub-constant error-probability low-degree test, and a subconstant error-probability pcp characterization of np. In Proceedings of the twenty-ninth annual ACM Symposium on Theory of Computing, pages 475–484, 1997.
  • [Tre14] Luca Trevisan. Inapproximability of combinatorial optimization problems. Paradigms of Combinatorial Optimization: Problems and New Approaches, pages 381–434, 2014.
  • [TW23] Theophile Thiery and Justin Ward. An improved approximation for maximum weighted k-set packing. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 1138–1162. SIAM, 2023.

Appendix A Proofs of Lemmas 2.3 and  2.4

In this section we prove Lemmas  2.3 and  8.4. The proofs of these lemmas requires tools from [EKL23, EKL24] regarding Fourier analysis over the Bilinear Scheme.

A.1 Fourier Analysis over the Bilinear Scheme

The key to proving Lemma 2.3 is a level-d inequality for indicator functions on the Bilinear Scheme due to Evra, Kindler, and Lifshitz [EKL24]. In order to use the result of [EKL24], however, we first give some necessary background for Fourier analysis over the Bilinear Scheme, and describe the analogues of zoom-ins, zoom-outs, and pseudorandomness. The latter is done in [EKL23, EKL24]. After doing so, we must then find a suitable map from the Grassmann graph to the Bilinear Scheme that (1) preserves the edges of our original bipartite inclusion graph between 22\ell-dimensional and 2(1δ)2(1-\delta)\ell subspaces, and (2) maps zoom-ins and zoom-outs in the Grassmann graph to their analogues over the Bilinear Scheme.

The Bilinear Scheme:

Let 𝔽qn×2\mathbb{F}_{q}^{n\times 2\ell} be the set of n×2n\times 2\ell matrices over 𝔽q\mathbb{F}_{q}. One can define a graph over 𝔽qn×2\mathbb{F}_{q}^{n\times 2\ell} that is similar to the Grassmann graphs by calling M1,M2𝔽qn×2M_{1},M_{2}\in\mathbb{F}_{q}^{n\times 2\ell} adjacent if dim(ker(M1M2))s\dim(\ker(M_{1}-M_{2}))\leqslant s for some s2s\leqslant 2\ell. A graph of this form are often referred to as the Bilinear Scheme. For our purposes, we will need to work with a bipartite version of this graph between 𝔽q(n2)×2\mathbb{F}_{q}^{(n-2\ell)\times 2\ell} and 𝔽q(n2(1δ))×2(1δ)\mathbb{F}_{q}^{(n-2(1-\delta)\ell)\times 2(1-\delta)\ell}. We equip the space L2(𝔽qn×2)L_{2}(\mathbb{F}_{q}^{n\times 2\ell}) with the following inner product:

F,G=𝔼M𝔽qn×2[F(M)G(M)¯],\langle F,G\rangle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}[F(M)\overline{G(M)}],

where the distribution taken over MM is uniform. Let ω\omega be a primitive ppth root of unity, where recall pp is the characteristic of 𝔽q\mathbb{F}_{q}. For s𝔽qns\in\mathbb{F}_{q}^{n} and x𝔽qnx\in\mathbb{F}_{q}^{n}, let χs(x)=ωTr(sx)\chi_{s}(x)=\omega^{\operatorname{Tr}(s\cdot x)} where Tr:𝔽q𝔽p\operatorname{Tr}\colon\mathbb{F}_{q}\to\mathbb{F}_{p} is the trace map. Then, the characters, χS:𝔽qn×2\chi_{S}:\mathbb{F}_{q}^{n\times 2\ell}\xrightarrow[]{}\mathbb{C} over all S=(s1,,s2)𝔽qn×2S=(s_{1},\ldots,s_{2\ell})\in\mathbb{F}_{q}^{n\times 2\ell}, given by

χS(x1,,x2)=i=12χsi(xi)=ωi=12Tr(sixi)\chi_{S}(x_{1},\ldots,x_{2\ell})=\prod_{i=1}^{2\ell}\chi_{s_{i}}(x_{i})=\omega^{\sum_{i=1}^{2\ell}\operatorname{Tr}(s_{i}\cdot x_{i})}

form an orthonormal basis of L2(𝔽qn×2)L_{2}(\mathbb{F}_{q}^{n\times 2\ell}). As a result, any FL2(𝔽qn×2)F\in L_{2}(\mathbb{F}_{q}^{n\times 2\ell}) can be expressed as,

F=S𝔽qn×2F^(S)χS,F=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell}}\widehat{F}(S)\chi_{S},

where F^(S)=F,χS\widehat{F}(S)=\langle F,\chi_{S}\rangle. The level dd component of FF is given by

F=d=S:rank(S)=dF^(S)χS.F^{=d}=\sum_{S:\;\operatorname{rank}(S)=d}\widehat{F}(S)\chi_{S}.

If a function FF only consists of components up to level dd, i.e. F^(S)=0\widehat{F}(S)=0 for all rank(S)>d\operatorname{rank}(S)>d, then we say FF is of degree dd.

We now describe the analogues of zoom-ins and zoom-outs on 𝔽qn×2\mathbb{F}_{q}^{n\times 2\ell}. We also define the analogous notion of (r,ε)(r,\varepsilon)-pseudo-randomness for Boolean functions over 𝔽qn×2\mathbb{F}_{q}^{n\times 2\ell}, and we begin by defining the analog of zoom-ins.

Definition A.1.

A zoom-in of dimension dd over 𝔽qn×2\mathbb{F}_{q}^{n\times 2\ell} is given by dd-pairs of vectors (u1,v1),,(ur,vr)(u_{1},v_{1}),\ldots,(u_{r},v_{r}) where each ui𝔽q2u_{i}\in\mathbb{F}_{q}^{2\ell} and each vi𝔽qnv_{i}\in\mathbb{F}_{q}^{n}. Let U𝔽q2×dU\in\mathbb{F}_{q}^{2\ell\times d} and V𝔽qn×dV\in\mathbb{F}_{q}^{n\times d} denote the matrices whose iith columns are uiu_{i} and viv_{i} respectively. Then the zoom-in on (U,V)(U,V) is the set of M𝔽qn×2M\in\mathbb{F}_{q}^{n\times 2\ell} such that MU=VMU=V, or equivalently, Mui=viMu_{i}=v_{i} for 1id1\leqslant i\leqslant d.

Next, we define the analog of zoom-outs.

Definition A.2.

A zoom-out of dimension dd is defined similarly, except by multiplication on the left. Given X𝔽qd×nX\in\mathbb{F}_{q}^{d\times n} and Y𝔽qd×2Y\in\mathbb{F}_{q}^{d\times 2\ell}, whose rows are given by xix_{i} and yiy_{i} respectively, the zoom-out (X,Y)(X,Y) is the M𝔽qn×2M\in\mathbb{F}_{q}^{n\times 2\ell} such that XM=YXM=Y, or equivalently, xiM=yix_{i}M=y_{i} for 1id1\leqslant i\leqslant d.

Let 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] denote the intersections of the zoom-in on (U,V)(U,V) and the zoom-out on (X,Y)(X,Y). The codimension of 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] is the sum of the number of columns of UU and the number of rows of XX, which we will denote by dim(U)\dim(U) and codim(X)\operatorname{codim}(X). For a zoom-in and zoom-out pair and a Boolean function FF, we define F(U,V),(X,Y):𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{0,1}F_{(U,V),(X,Y)}:{\sf Zoom}[(U,V),(X,Y)]\xrightarrow[]{}\{0,1\} to be the restriction of FF which is given as

F(U,V),(X,Y)(M)=F(M)forM𝖹𝗈𝗈𝗆[(U,V),(X,Y)].F_{(U,V),(X,Y)}(M)=F(M)\quad\text{for}\quad M\in{\sf Zoom}[(U,V),(X,Y)].

When dim(U)+codim(X)=d\dim(U)+\operatorname{codim}(X)=d, we say that the restriction is of size dd. We define (d,ε)(d,\varepsilon)-pseudo-randomness in terms of the L2L_{2}-norms of restrictions of FF of size dd. Here and throughout, when we consider restricted functions, the underlying measure is the uniform measure over the corresponding zoom-in and zoom-out set 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)].

Definition A.3.

We say that an indicator function FL2(𝔽qn×2)F\in L_{2}(\mathbb{F}_{q}^{n\times 2\ell}) is (d,ε)(d,\varepsilon)-pseudorandom if for all zoom-in zoom-out combinations 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] such that dim(U)+codim(X)=d\dim(U)+\operatorname{codim}(X)=d, we have

F(U,V),(X,Y)22ε.\left\lVert F_{(U,V),(X,Y)}\right\rVert_{2}^{2}\leqslant\varepsilon.

We note that for Boolean functions FF, F(U,V),(X,Y)22=𝔼M𝖹𝗈𝗈𝗆[(U,V),(X,Y)][F(M)]\left\lVert F_{(U,V),(X,Y)}\right\rVert_{2}^{2}=\mathop{\mathbb{E}}_{M\in{\sf Zoom}[(U,V),(X,Y)]}[F(M)], and hence the definition above generalizes the definition we have for Boolean functions.

Definition A.4.

We say that an indicator function FL2(𝔽qn×2)F\in L_{2}(\mathbb{F}_{q}^{n\times 2\ell}) is (d,ε,t)(d,\varepsilon,t)-pseudo-random if for all 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] such that dim(U)+codim(X)=d\dim(U)+\operatorname{codim}(X)=d, we have,

F(U,V),(X,Y)t/(t1)=(𝔼M𝖹𝗈𝗈𝗆[(U,V),(X,Y)][|F(M)|tt1])t1tε.\left\lVert F_{(U,V),(X,Y)}\right\rVert_{t/(t-1)}=\left(\mathop{\mathbb{E}}_{M\in{\sf Zoom}[(U,V),(X,Y)]}\left[|F(M)|^{\frac{t}{t-1}}\right]\right)^{\frac{t-1}{t}}\leqslant\varepsilon.

The following result is a combination of two results form [EKL24]. Roughly speaking, it states that if a Boolean function FF is (r,ε)(r,\varepsilon) pseudo-random, then its degree dd parts are (r,Cq,dε2)(r,C_{q,d}\varepsilon^{2}) pseudo-random for drd\leqslant r.

Lemma A.5.

[EKL24, Theorem 5.5 + Proposition 3.6] Let t4t\geqslant 4 be a power of 22 and let F:𝔽qn×2{0,1}F:\mathbb{F}_{q}^{n\times 2\ell}\to\{0,1\} be a function that is (d,ε,t)(d,\varepsilon,t)-pseudo-random. Then F=dF^{=d} is (r,q10dr+500d2tε2)(r,q^{10dr+500d^{2}t}\varepsilon^{2})-pseudo-random for all rdr\geqslant d.

Proof.

This lemma does not actually appear in [EKL24], but it is easy to derive by combining Theorem 5.5 with Proposition 3.6 therein. In [EKL23, EKL24], the authors introduce an additional notion of generalized influences and having small generalized influences. We refrain from defining these notions explicitly as it is slightly cumbersome, but roughly speaking, one defines a Laplacian for each zoom-in, zoom-out combination, so that having (d,ε)(d,\varepsilon) small generalized influences means that upon applying these Laplacians on FF, the 22-norm squared of the resulting function never exceeds ε\varepsilon.

With this notion in hand, if a function FF is (d,ε,t)(d,\varepsilon,t)-pseudo-random, then by [EKL24, Theorem 5.5] we get that F=dF^{=d} has (d,q500d2tε2)(d,q^{500d^{2}t}\varepsilon^{2})-small generalized influences. Applying [EKL24, Proposition 3.6] then implies that F=dF^{=d} is (r,q10drq500d2tε2)(r,q^{10dr}\cdot q^{500d^{2}t}\varepsilon^{2})-pseudo-random for any rdr\geqslant d, which is the desired result ∎

Lastly, we need the following global hypercontrativity result also due to [EKL24].999We remark that earlier results [EKL23] showed similar statement for 44-norms, i.e. the case that t=4t=4, and the result below follows by a form of induction on tt. That is, one starts with FF and concludes via applying the case t=4t=4 that the function and it F2F^{2} is (d,Cq,dε)(d,C_{q,d}\varepsilon) pseudo-random. Then one apply the case t=4t=4 on F2F^{2} to conclude that F4F^{4} is (d,Cq,dε)(d,C_{q,d}^{\prime}\varepsilon) pseudo-random and so on.

Theorem A.6.

[EKL24, Theorem 1.13] Let t4t\geqslant 4 be a power of 22 and let FL2(𝔽qn×2)F\in L_{2}(\mathbb{F}_{q}^{n\times 2\ell}) be a function of degree dd that is (d,ε)(d,\varepsilon)-pseudo-random. Then,

Fttq200d2t2F22εt/21.\left\lVert F\right\rVert_{t}^{t}\leqslant q^{200d^{2}t^{2}}\left\lVert F\right\rVert_{2}^{2}\varepsilon^{t/2-1}.

Combining Lemma A.5 and Theorem A.6, we arrive at the following result which bounds the tt-norm of the level dd component of pseudo-random indicator functions. This result will be the key to showing an analogue of Lemma 2.3 over the Bilinear Scheme.

Theorem A.7.

Let t4t\geqslant 4 be a power of 22. Then if F:𝔽qn×2{0,1}F:\mathbb{F}_{q}^{n\times 2\ell}\xrightarrow[]{}\{0,1\} is (r,ε)(r,\varepsilon)-pseudo-random, we have

F=dtq500d2tεt1t\left\lVert F^{=d}\right\rVert_{t}\leqslant q^{500d^{2}t}\varepsilon^{\frac{t-1}{t}}

for all drd\leqslant r.

Proof.

Suppose FF is (r,ε)(r,\varepsilon)-pseudo-random, let t4t\geqslant 4 be a power of 22, and fix a drd\leqslant r. Since drd\leqslant r, we also have that FF is (d,ε)(d,\varepsilon)-peudorandom. Therefore for any size dd restriction of FF, F(U,V),(X,Y)F_{(U,V),(X,Y)} ,we have,

F(U,V),(X,Y)t/(t1)=(F(U,V),(X,Y)22)t1tεt1t.\left\lVert F_{(U,V),(X,Y)}\right\rVert_{t/(t-1)}=\left(\left\lVert F_{(U,V),(X,Y)}\right\rVert_{2}^{2}\right)^{\frac{t-1}{t}}\leqslant\varepsilon^{\frac{t-1}{t}}.

Thus, FF is (d,εt1t,t)(d,\varepsilon^{\frac{t-1}{t}},t)-pseudo-random, and by Lemma A.5 it follows that F=dF^{=d} is (d,q10d2+500d2tε2t2t)(d,q^{10d^{2}+500d^{2}t}\varepsilon^{\frac{2t-2}{t}})-pseudo-random. Clearly, F=dF^{=d} is degree dd, so applying Theorem A.6 we get that,

F=dttq200d2t2F=d22(q10d2+500d2tε2t2t)t/21q500d2t2εt1,\left\lVert F^{=d}\right\rVert_{t}^{t}\leqslant q^{200d^{2}t^{2}}\left\lVert F^{=d}\right\rVert_{2}^{2}\left(q^{10d^{2}+500d^{2}t}\varepsilon^{\frac{2t-2}{t}}\right)^{t/2-1}\leqslant q^{500d^{2}t^{2}}\varepsilon^{t-1},

where we also use the fact that F=d22F22ε\left\lVert F^{=d}\right\rVert_{2}^{2}\leqslant\left\lVert F\right\rVert_{2}^{2}\leqslant\varepsilon because FF is an (r,ε)(r,\varepsilon)-pseudorandom Boolean function. Taking the tt-th root of the above inequality completes the proof. ∎

A.2 An Analog of Lemmas 2.3 for the Bilinear Scheme

With Theorem A.7 in hand we can show an analogue of Lemma 2.3 for basis invariant functions over the Bilinear Scheme. To do so, we first define what we mean by basis invariant functions, then present an analogue of the adjacency operator 𝒯\mathcal{T} (which is originally defined for functions over subspaces) over the Bilinear Scheme, which we denote by 𝒯\mathcal{T}^{\prime}, and finally show that the previously described characters are eigenoperators of 𝒯𝒯\mathcal{T}^{\prime*}\circ\mathcal{T}^{\prime}, where 𝒯\mathcal{T}^{\prime*} is an operator that acts as the adjoint of 𝒯\mathcal{T}^{\prime} on basis invariant functions.

For a function FL2(𝔽qn×2)F\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right), say that FF is basis invariant if F(M)=F(MA)F(M)=F(MA) for any full rank A𝔽q2×2A\in\mathbb{F}_{q}^{2\ell\times 2\ell}. We first show that the level dd component of a basis invariant function is also basis invariant. The following identity regarding the characters will be useful.

Lemma A.8.

For any S=(s1,,s)𝔽qn×S=(s_{1},\ldots,s_{\ell^{\prime}})\in\mathbb{F}_{q}^{n\times\ell^{\prime}}, any M𝔽qn×2M\in\mathbb{F}_{q}^{n\times 2\ell}, and any matrix A𝔽q2×A\in\mathbb{F}_{q}^{2\ell\times\ell^{\prime}} we have,

χS(MA)=χSAT(M).\chi_{S}(MA)=\chi_{SA^{T}}(M).
Proof.

Letting v1,,v2v_{1},\ldots,v_{2\ell} denote the columns of MM and ai,ja_{i,j} denote the entries of AA, we have,

χS(MA)=ωTri=1si(j=12vjaj,i)=ωi=1Tr(vi(j=12sjai,j))=χSAT(M).\chi_{S}(MA)=\omega^{\operatorname{Tr}\sum_{i=1}^{\ell^{\prime}}s_{i}\cdot\left(\sum_{j=1}^{2\ell}v_{j}a_{j,i}\right)}=\omega^{\sum_{i=1}^{\ell^{\prime}}\operatorname{Tr}(v_{i}\cdot\left(\sum_{j=1}^{2\ell}s_{j}a_{i,j}\right))}=\chi_{SA^{T}}(M).\qed
Lemma A.9.

Let S=[s1,,s2]𝔽qn×2S=[s_{1},\ldots,s_{2\ell}]\in\mathbb{F}_{q}^{n\times 2\ell} and let FL2(𝔽qn×2)F\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) be basis invariant. Then for any A𝔽q2×2A\in\mathbb{F}_{q}^{2\ell\times 2\ell} that is full rank, we have F^(SA)=F^(S)\widehat{F}(SA)=\widehat{F}(S).

Proof.

For any matrix full rank B𝔽q2×2B\in\mathbb{F}_{q}^{2\ell\times 2\ell} we have

F^(S)\displaystyle\widehat{F}(S) =𝔼M𝔽qn×2[χS(M)F(M)]\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}\left[\chi_{S}(M)F(M)\right]
=𝔼M𝔽qn×2[χS(M)F(MB1)]\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}\left[\chi_{S}(M)F(MB^{-1})\right]
=𝔼M𝔽qn×2[χS(MB)F(M)]\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}\left[\chi_{S}(MB)F(M)\right]
=EM𝔽qn×2[χSBT(M)F(M)]\displaystyle=E_{M\in\mathbb{F}_{q}^{n\times 2\ell}}[\chi_{SB^{T}}(M)F(M)]
=F^(SBT),\displaystyle=\widehat{F}(SB^{T}),

where we use that FF is basis invariant in the third transition and Lemma A.8 in the fourth transition. Setting B=ATB=A^{T} gives the result. ∎

Using Lemma A.9, we can show that the level dd component of a basis invariant function is also basis invariant.

Lemma A.10.

If FL2(𝔽qn×2)F\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) is basis invariant, then F=dF^{=d} is basis invariant as well for any dd.

Proof.

Fix any M𝔽qn×2M\in\mathbb{F}_{q}^{n\times 2\ell} and A𝔽q2×2A\in\mathbb{F}_{q}^{2\ell\times 2\ell} full rank. We have

F=d(MA)\displaystyle F^{=d}(MA) =S𝔽qn×2,rank(S)=dF^(S)χS(MA)\displaystyle=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell},\operatorname{rank}(S)=d}\widehat{F}(S)\chi_{S}(MA)
=S𝔽qn×2,rank(S)=dF^(S)χSAT(M)\displaystyle=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell},\operatorname{rank}(S)=d}\widehat{F}(S)\chi_{SA^{T}}(M)
=S𝔽qn×2,rank(S)=dF^(S(AT)1)χS(M)\displaystyle=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell},\operatorname{rank}(S)=d}\widehat{F}(S(A^{T})^{-1})\chi_{S}(M)
=S𝔽qn×2,rank(S)=dF^(S)χS(M)\displaystyle=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell},\operatorname{rank}(S)=d}\widehat{F}(S)\chi_{S}(M)
=F=d(M),\displaystyle=F^{=d}(M),

where we use Lemma A.8 in the second transition, and Lemma A.9 in the fourth transition. ∎

We now define the following two operators which will be the analogues of 𝒯\mathcal{T} and 𝒯\mathcal{T}^{*} over the bilinear scheme. The first is 𝒯:L2(𝔽qn×2)L2(𝔽qn×2(1δ))\mathcal{T}^{\prime}:L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right)\xrightarrow{}L_{2}\left(\mathbb{F}_{q}^{n\times 2(1-\delta)\ell}\right), given by:

𝒯F(M)=𝔼v1,,vδ[F([M,v1,,v2δ)].\mathcal{T}^{\prime}F(M^{\prime})=\mathop{\mathbb{E}}_{v_{1},\ldots,v_{\delta\ell}}[F\left([M^{\prime},v_{1},\ldots,v_{2\delta\ell}\right)].

In words, the operator 𝒯\mathcal{T}^{\prime} averages over extensions of the matrix MM to an n×2n\times 2\ell matrix by adding to it 2δ2\delta\ell random columns. The next is 𝒯:L2(𝔽qn×2(1δ))L2(𝔽qn×2)\mathcal{T}^{\prime*}:L_{2}\left(\mathbb{F}_{q}^{n\times 2(1-\delta)\ell}\right)\xrightarrow{}L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) given by:

𝒯G(M)=𝔼A𝔽q2×2(1δ)[G(MA)|rank(A)=2(1δ)].\mathcal{T}^{\prime*}G(M)=\mathop{\mathbb{E}}_{A\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell}}[G(MA)\;|\;\operatorname{rank}(A)=2(1-\delta)\ell].

Strictly speaking, 𝒯\mathcal{T}^{\prime*} is not the adjoint of 𝒯\mathcal{T}^{\prime}; however, for the case where FF is basis invariant, 𝒯\mathcal{T}^{\prime*} acts as the adjoint of 𝒯\mathcal{T}^{\prime} in the following sense.

Lemma A.11.

For FL2(𝔽qn×2)F\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) that is basis invariant and GL2(𝔽qn×2(1δ))G\in L_{2}\left(\mathbb{F}_{q}^{n\times 2(1-\delta)\ell}\right), we have

𝒯F,G=F,𝒯G.\langle\mathcal{T}^{\prime}F,G\rangle=\langle F,\mathcal{T}^{\prime*}G\rangle.
Proof.

Let J𝔽q2×2(1δ)J\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell} be the matrix whose restriction to the first 2(1δ)2(1-\delta)\ell rows is the identity matrix I2(1δ)×2(1δ)I_{2(1-\delta)\ell\times 2(1-\delta)\ell} and whose remaining rows are all 0. We have

𝒯F,G\displaystyle\langle\mathcal{T}^{\prime}F,G\rangle =𝔼M𝔽qn×2(1δ),v1,,v2δ𝔽qn[F([M,v1,,v2δ])G(M)¯]\displaystyle=\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times 2(1-\delta)\ell},v_{1},\ldots,v_{2\delta\ell}\in\mathbb{F}_{q}^{n}}\left[F^{\prime}\left([M^{\prime},v_{1},\ldots,v_{2\delta\ell}]\right)\cdot\overline{G(M^{\prime})}\right]
=𝔼M𝔽qn×2(1δ),vi𝔽qn,A𝔽q2×2[F([M,v1,,v2δ]A)G(M)¯|rank(A)=2]\displaystyle=\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times 2(1-\delta)\ell},v_{i}\in\mathbb{F}_{q}^{n},A\in\mathbb{F}_{q}^{2\ell\times 2\ell}}\left[F^{\prime}\left([M^{\prime},v_{1},\ldots,v_{2\delta\ell}]A\right)\cdot\overline{G(M^{\prime})}\;|\;\operatorname{rank}(A)=2\ell\right]
=𝔼M𝔽qn×2,A𝔽q2×2[F(M)G(MA1J)¯|rank(A)=2],\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell},A\in\mathbb{F}_{q}^{2\ell\times 2\ell}}[F(M)\cdot\overline{G(MA^{-1}J)}\;|\;\operatorname{rank}(A)=2\ell],

where in the second transition we used the fact that FF^{\prime} is basis invariant, and in the third one we made a change of variables M=[M,v1,,v2δ]AM=[M^{\prime},v_{1},\ldots,v_{2\delta\ell}]A. Now note that A1JA^{-1}J is the matrix A1A^{-1} restricted to its first 2(1δ)2(1-\delta)\ell columns and hence in the final distribution, A1JA^{-1}J is a uniformly random matrix in 𝔽qn×2(1δ)\mathbb{F}_{q}^{n\times 2(1-\delta)\ell} with rank 2(1δ)2(1-\delta)\ell. It follows that,

𝒯F,G=𝔼M𝔽qn×2,A𝔽q2×2(1δ)[F(M)G(MA)¯|rank(A)=2(1δ)]=F,𝒯G.\langle\mathcal{T}^{\prime}F,G\rangle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell},A\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell}}[F(M)\cdot\overline{G(MA)}\;|\;\operatorname{rank}(A)=2(1-\delta)\ell]=\langle F,\mathcal{T}^{\prime*}G\rangle.\qed

We will want to understand the operator 𝒯𝒯\mathcal{T}^{\prime*}\mathcal{T}^{\prime}, and towards this end we define the operator

ΦF(M)=𝔼B𝔽qn×2δC𝔽q2δ×2rank(C)=2δ[F(M+BC)].\Phi F(M)=\mathop{\mathbb{E}}_{\begin{subarray}{c}B\in\mathbb{F}_{q}^{n\times 2\delta\ell}\\ C\in\mathbb{F}_{q}^{2\delta\ell\times 2\ell}\\ \operatorname{rank}(C)=2\delta\ell\end{subarray}}\left[F(M+BC)\right].

The reason for introducing Φ\Phi is that, as the following lemma shows, it acts the same as 𝒯𝒯\mathcal{T}^{\prime*}\mathcal{T}^{\prime} on basis invariant functions, but is easier to work with. This is due to the reason it is an averaging operator with respect to some Cayley graph over 𝔽qn×2\mathbb{F}_{q}^{n\times 2\ell}, and therefore each character χS\chi_{S} is an eigenvector of Φ\Phi and the eigenvalues have an explicit formula. These facts are shown in the next two lemmas respectively.

Lemma A.12.

If FL2(𝔽qn×2)F\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) is basis invariant, then 𝒯𝒯F=ΦF{\mathcal{T}^{\prime}}^{*}\mathcal{T}^{\prime}F=\Phi F.

Proof.

By definitions

𝒯𝒯F(M)=𝔼R𝔽q2×2(1δ),v1,,v2δ𝔽qn[F([MR,v1,,v2δ]|rank(R)=2(1δ))].\mathcal{T}^{\prime*}\mathcal{T}^{\prime}F(M)=\mathop{\mathbb{E}}_{\begin{subarray}{c}R^{\prime}\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell},\\ v_{1},\ldots,v_{2\delta\ell}\in\mathbb{F}_{q}^{n}\end{subarray}}[F^{\prime}\left([MR^{\prime},v_{1},\ldots,v_{2\delta\ell}]~{}|~{}\operatorname{rank}(R^{\prime})=2(1-\delta)\ell\right)].

We can also view M=[MR,v1,,v2δ]M^{\prime}=[MR^{\prime},v_{1},\ldots,v_{2\delta\ell}] as being sampled as follows. Choose R𝔽q2×2(1δ)R^{\prime}\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell} with linearly independent columns, extend RR^{\prime} to a matrix R𝔽q2×2R\in\mathbb{F}_{q}^{2\ell\times 2\ell} with linearly independent columns randomly by adding 2δ2\delta\ell columns on the right, sample a random matrix [0,,0,w1,,w2δ]𝔽qn×2[0,\dots,0,w_{1},\ldots,w_{2\delta\ell}]\in\mathbb{F}_{q}^{n\times 2\ell}, and output,

M=MR+[0,,0,w1,,w2δ].M^{\prime}=MR+[0,\dots,0,w_{1},\ldots,w_{2\delta\ell}].

Furthermore, under this distribution, it is clear that R𝔽q2×2R\in\mathbb{F}_{q}^{2\ell\times 2\ell} is a uniformly random matrix with linearly independent columns. Therefore,

𝒯𝒯F(M)\displaystyle\mathcal{T}^{\prime*}\mathcal{T}^{\prime}F(M) =𝔼R𝔽q2×2,w1,,w2δ𝔽qn[F(MR+[0,,0,w1,,w2δ]]\displaystyle=\mathop{\mathbb{E}}_{\begin{subarray}{c}R\in\mathbb{F}_{q}^{2\ell\times 2\ell},\\ w_{1},\ldots,w_{2\delta\ell}\in\mathbb{F}_{q}^{n}\end{subarray}}[F(MR+[0,\dots,0,w_{1},\ldots,w_{2\delta\ell}]]
=𝔼R𝔽q2×2,w1,,w2δ𝔽qn[F(M+[0,,0,w1,,w2δ]R1)],\displaystyle=\mathop{\mathbb{E}}_{\begin{subarray}{c}R\in\mathbb{F}_{q}^{2\ell\times 2\ell},\\ w_{1},\ldots,w_{2\delta\ell}\in\mathbb{F}_{q}^{n}\end{subarray}}[F(M+[0,\dots,0,w_{1},\ldots,w_{2\delta\ell}]R^{-1})],

where we are using the fact that FF is basis invariant and RR is invertible. In the last expectation, note that the distribution over [0,,0,w1,,w2δ]R1[0,\dots,0,w_{1},\ldots,w_{2\delta\ell}]R^{-1} is the same as that over BCBC where B𝔽qn×2δB\in\mathbb{F}_{q}^{n\times 2\delta\ell} is uniformly random, and C𝔽q2δ×2C\in\mathbb{F}_{q}^{2\delta\ell\times 2\ell} is uniformly random conditioned on having linearly independent rows. More precisely, it is equal to BCBC where B=[w1,,w2δ]B=[w_{1},\ldots,w_{2\delta\ell}], and CC is the last 2δ2\delta\ell rows of R1R^{-1}. It follows that

𝒯𝒯F(M)=𝔼B𝔽qn×2δ,C𝔽q2δ×2[F(M+BC)|rank(C)=2δ].\mathcal{T}^{\prime*}\mathcal{T}^{\prime}F(M)=\mathop{\mathbb{E}}_{B\in\mathbb{F}_{q}^{n\times 2\delta\ell},C\in\mathbb{F}_{q}^{2\delta\ell\times 2\ell}}\left[F(M+BC)\;|\;\operatorname{rank}(C)=2\delta\ell\right].\qed

The following lemma gives upper bound on the eigenvalues of Φ\Phi.

Lemma A.13.

Suppose that rank(S)=t\operatorname{rank}(S)=t. If t=0t=0, then χS\chi_{S} is an eigenvector of Φ\Phi of eigenvalue 11. Otherwise, if t>0t>0, χS\chi_{S} is an eigenvector of Φ\Phi of eigenvalue which is at most 3qtn+qt(2δ1)3q^{t-n}+q^{-t(2\delta\ell-1)} in absolute value.

Proof.

Fix SS. We argued earlier that χS\chi_{S} is an eigenvector of Φ\Phi, and we denote the corresponding eigevalue by λ=ΦχS(0)\lambda=\Phi\chi_{S}(0). If t=0t=0 the statement is clear, so we assume that t>0t>0 henceforth.

Find A𝔽q2×2A\in\mathbb{F}_{q}^{2\ell\times 2\ell} of full rank so that SAT=(v1,,vt,0,0,,0)SA^{T}=(v_{1},\ldots,v_{t},0,0,\ldots,0) where v1,,vtv_{1},\ldots,v_{t} are linearly independent. Thus, as the distribution of CC is invariant under multiplying by ATA^{T} from the right, we get that

λ=ΦχS(0)=𝔼B,C[χS(BCAt)|rank(C)=2δ]=𝔼B,C[χSA(BC)|rank(C)=2δ],\lambda=\Phi\chi_{S}(0)=\mathop{\mathbb{E}}_{B,C}[\chi_{S}(BCA^{t})~{}|~{}\operatorname{rank}(C)=2\delta\ell]=\mathop{\mathbb{E}}_{B,C}[\chi_{SA}(BC)~{}|~{}\operatorname{rank}(C)=2\delta\ell],

and we may assume that S=(v1,,vt,0,,0)S=(v_{1},\ldots,v_{t},0,\ldots,0) for linearly independent v1,,vtv_{1},\ldots,v_{t} to begin with. Applying symmetry again, we conclude that

λ=𝔼v1,,vtlinearly independent[Φχ(v1,,vt,0)(0)]=𝔼v1,,vtlinearly independent[𝔼B,Cωi=1tTr(vi𝖼𝗈𝗅i(BC))],\lambda=\mathop{\mathbb{E}}_{\begin{subarray}{c}v_{1},\ldots,v_{t}\\ \text{linearly independent}\end{subarray}}\left[\Phi\chi_{(v_{1},\ldots,v_{t},\vec{0})}(0)\right]=\mathop{\mathbb{E}}_{\begin{subarray}{c}v_{1},\ldots,v_{t}\\ \text{linearly independent}\end{subarray}}\left[\mathop{\mathbb{E}}_{B,C}\omega^{\sum\limits_{i=1}^{t}\operatorname{Tr}(v_{i}\cdot{\sf col}_{i}(BC))}\right],

and interchanging the order of expectations we get that

λ=𝔼B,C[𝔼v1,,vtlinearly independentωi=1tTr(vi𝖼𝗈𝗅i(BC))],\lambda=\mathop{\mathbb{E}}_{B,C}\left[\mathop{\mathbb{E}}_{\begin{subarray}{c}v_{1},\ldots,v_{t}\\ \text{linearly independent}\end{subarray}}\omega^{\sum\limits_{i=1}^{t}\operatorname{Tr}(v_{i}\cdot{\sf col}_{i}(BC))}\right],

Denote wi=𝖼𝗈𝗅i(BC)w_{i}={\sf col}_{i}(BC), and inspect these vectors.

Claim A.14.

If wi0w_{i}\neq 0 for some ii, then

|𝔼v1,,vtlinearly independent[ωi=1tTr(vi𝖼𝗈𝗅i(BC))]|2qtn.\left|\mathop{\mathbb{E}}_{\begin{subarray}{c}v_{1},\ldots,v_{t}\\ \text{linearly independent}\end{subarray}}\left[\omega^{\sum\limits_{i=1}^{t}\operatorname{Tr}(v_{i}\cdot{\sf col}_{i}(BC))}\right]\right|\leqslant 2q^{t-n}.
Proof.

We first claim that if v1,,vtv_{1},\ldots,v_{t} are chosen uniformly, then the left hand side is 0, or equivalently

𝔼v1,,vtuniform[ωTr(i=1tviwi)]=0.\mathop{\mathbb{E}}_{\begin{subarray}{c}v_{1},\ldots,v_{t}\\ \text{uniform}\end{subarray}}\left[\omega^{\operatorname{Tr}\left(\sum\limits_{i=1}^{t}v_{i}\cdot w_{i}\right)}\right]=0.

To see this, it suffices to show that i=1tviwi\sum_{i=1}^{t}v_{i}\cdot w_{i} takes every value in 𝔽q\mathbb{F}_{q} with equal probability, and we focus on showing this. Fix ii such that wi0w_{i}\neq 0 and suppose the jjth entry, wi,jw_{i,j} is nonzero. We can fix all entries of the v1,,vtv_{1},\ldots,v_{t} uniformly except for vi,jv_{i,j}, and then for each α𝔽q\alpha\in\mathbb{F}_{q}, there is exactly one choice of vi,jv_{i,j} that will result in i=1tviwi=α\sum_{i=1}^{t}v_{i}\cdot w_{i}=\alpha.

Thus, if we took the distribution over v1,,vtv_{1},\ldots,v_{t} to be uniformly and independently chosen, then the magnitude of the above expectation would be 0. Hence, we conclude that the above expectation is at most twice the probability randomly chosen v1,,vtv_{1},\ldots,v_{t} are not linearly independent, which is at most qtnq^{t-n}. ∎

By Claim A.14 we conclude that λ2qtn+PrB,C[wi=0i=1,,t]\lambda\leqslant 2q^{t-n}+\Pr_{B,C}[w_{i}=0~{}\forall i=1,\ldots,t], and we next bound this probability. Recalling the definition of wiw_{i}, we have that

wi=j=12δC(j,i)𝖼𝗈𝗅j(B).w_{i}=\sum\limits_{j=1}^{2\delta\ell}C(j,i){\sf col}_{j}(B).

Consider the 2δ×t2\delta\ell\times t minor of CC and call it CC^{\prime}. First we upper bound the probability that rank(C)=0\operatorname{rank}(C^{\prime})=0. Note that the distribution of CC is the same as of A|2δ×2A|_{2\delta\ell\times 2\ell} where A𝔽q2×2A\in\mathbb{F}_{q}^{2\ell\times 2\ell} is a random invertible matrix. Thus, CC^{\prime} has the same distribution as of A|2δ×tA|_{2\delta\ell\times t}, and the probability that C=0C^{\prime}=0 is at most

q22δq21q22δq2qq22δq2qt1qt(2δ1).\frac{q^{2\ell-2\delta\ell}}{q^{2\ell}-1}\cdot\frac{q^{2\ell-2\delta\ell}}{q^{2\ell}-q}\cdots\frac{q^{2\ell-2\delta\ell}}{q^{2\ell}-q^{t-1}}\leqslant q^{-t(2\delta\ell-1)}.

It remains to bound the probability that wiw_{i} are all 0 in the case that rank(C)1\operatorname{rank}(C^{\prime})\geqslant 1. In this case, assume without loss of generality that the first column of CC^{\prime} is non-zero. Thus, it follows that over the randomness of BB, the vector w1w_{1} is uniformly chosen from 𝔽qn\mathbb{F}_{q}^{n}, and so the probability it is the all 0 vector is at most qnq^{-n}. Combining, we get that λ3qtn+qt(2δ1)\lambda\leqslant 3q^{t-n}+q^{-t(2\delta\ell-1)}. ∎

Finally, using Lemma A.8 again, we can show that 𝒯\mathcal{T}^{\prime*} does not increase the level of a function,

𝒯χS(M)=𝔼A𝔽q2×2(1δ)[χS(MA)]=𝔼A[χSAT(M)],\mathcal{T}^{\prime*}\chi_{S}(M)=\mathop{\mathbb{E}}_{A\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell}}[\chi_{S}(MA)]=\mathop{\mathbb{E}}_{A}[\chi_{SA^{T}}(M)], (17)

and obtain a useful identity for decomposing inner products.

Lemma A.15.

Let FL2(𝔽qn×2)F\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) and GL2(𝔽qn×2(1δ))G\in L_{2}\left(\mathbb{F}_{q}^{n\times 2(1-\delta)\ell}\right). Then,

𝒯F=d,G=𝒯F=d,G=d.\langle\mathcal{T}^{\prime}F^{=d},G\rangle=\langle\mathcal{T}^{\prime}F^{=d},G^{=d}\rangle.

As a consequence we also have

F=d,𝒯G=𝒯F=d,𝒯G=d.\langle F^{=d},\mathcal{T}^{\prime*}G\rangle=\langle\mathcal{T}^{\prime}F^{=d},\mathcal{T}^{\prime*}G^{=d}\rangle.
Proof.

Using Equation (17), we have,

𝒯G=j(M)=S𝔽qn×2(1δ),rank(S)=jG^(S)𝔼A[χSAT(M)|rank(A)=2(1δ)].\displaystyle\mathcal{T}^{\prime*}G^{=j}(M)=\sum_{S\in\mathbb{F}_{q}^{n\times 2(1-\delta)\ell},\operatorname{rank}(S)=j}\widehat{G}(S)\mathop{\mathbb{E}}_{A}\left[\chi_{SA^{T}}(M)\;|\;\operatorname{rank}(A)=2(1-\delta)\ell\right].

Since rank(S)=j\operatorname{rank}(S)=j, it follows that the rank(SAT)\operatorname{rank}(SA^{T}) is at most jj, so it follows that for j<dj<d, we have F=d,𝒯G=j=0\langle F^{=d},\mathcal{T}^{*}G^{=j}\rangle=0. As a result,

𝒯F=d,G=F=d,𝒯G=j=d2(1δ)F=d,𝒯G.\langle\mathcal{T}^{\prime}F^{=d},G\rangle=\langle F^{=d},\mathcal{T}^{\prime*}G\rangle=\sum_{j=d}^{2(1-\delta)\ell}\langle F^{=d},\mathcal{T}^{\prime*}G\rangle. (18)

Next we have,

𝒯F=d(M)=S𝔽qn×2,rank(S)=dF^(S)χS(M)=S𝔽qn×2,rank(S)=dF^(S)χS(M),\mathcal{T}^{\prime}F^{=d}(M)=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell},\operatorname{rank}(S)=d}\widehat{F}(S)\chi_{S}(M^{\prime})=\sum_{S\in\mathbb{F}_{q}^{n\times 2\ell},\operatorname{rank}(S)=d}\widehat{F}(S)\chi_{S^{\prime}}(M^{\prime}),

where both MM^{\prime} and SS^{\prime} are obtained from MM by removing the last 2δ2\delta\ell columns. It follows that 𝒯F=d\mathcal{T}^{\prime}F^{=d} has level at most dd, so using (18) we get

𝒯F=d,G=j=d2(1δ)F=d,𝒯G=j=j=d2(1δ)𝒯F=d,G=j=𝒯F=d,G=d.\langle\mathcal{T}^{\prime}F^{=d},G\rangle=\sum_{j=d}^{2(1-\delta)\ell}\langle F^{=d},\mathcal{T}^{\prime*}G^{=j}\rangle=\sum_{j=d}^{2(1-\delta)\ell}\langle\mathcal{T}^{\prime}F^{=d},G^{=j}\rangle=\langle\mathcal{T}F^{=d},G^{=d}\rangle.\qed

We are now ready to state and prove an analog of Lemma 2.3 for basis invariant functions on the Bilinear scheme.

Lemma A.16.

Let FL2(𝔽qn×2)F\in L_{2}(\mathbb{F}_{q}^{n\times 2\ell}) and GL2(𝔽qn×2(1δ))G\in L_{2}(\mathbb{F}_{q}^{n\times 2(1-\delta)\ell}) be basis invariant indicator functions with 𝔼[F]=α,𝔼[G]=β\mathop{\mathbb{E}}[F]=\alpha,\mathop{\mathbb{E}}[G]=\beta. If FF is (r,ε)(r,\varepsilon) pseudo-random and basis invariant, then for all t4t\geqslant 4 that are powers of 22, we have

𝒯F,GqOt,r(1)β(t1)/tε2t/(2t1)+qrδαβ.\langle\mathcal{T}^{\prime}F,G\rangle\leqslant q^{O_{t,r}(1)}\beta^{(t-1)/t}\varepsilon^{2t/(2t-1)}+q^{-r\delta\ell}\sqrt{\alpha\beta}.
Proof.

Using the degree decomposition of FF and Lemma A.15, we can write

𝒯F,G=d=02𝒯F=d,G=d.\langle{\mathcal{T}^{\prime}F},{G}\rangle=\sum_{d=0}^{2\ell}\langle{\mathcal{T}^{\prime}F^{=d}},{G^{=d}}\rangle.

We first bound the contribution from terms in the summation with d>rd>r using Cauchy-Schwarz. For d>rd>r,

|𝒯F=d,G=d|2\displaystyle|\langle{\mathcal{T}^{\prime}F^{=d}},{G^{=d}}\rangle|^{2} 𝒯F=d22G=d22\displaystyle\leqslant\left\lVert\mathcal{T}^{\prime}F^{=d}\right\rVert_{2}^{2}\left\lVert G^{=d}\right\rVert_{2}^{2}
=G=d22F=d,𝒯𝒯F=d\displaystyle=\left\lVert G^{=d}\right\rVert_{2}^{2}\langle{F^{=d}},{\mathcal{T}^{\prime*}\mathcal{T}^{\prime}F^{=d}}\rangle
=G=d22F=d,ΦF=d\displaystyle=\left\lVert G^{=d}\right\rVert_{2}^{2}\langle{F^{=d}},{\Phi F^{=d}}\rangle
(q2dδ+3qdn)F=d22G=d22\displaystyle\leqslant\left(q^{-2d\delta\ell}+3q^{d-n}\right)\left\lVert F^{=d}\right\rVert_{2}^{2}\left\lVert G^{=d}\right\rVert_{2}^{2}
(q2dδ+3qdn)αβ,\displaystyle\leqslant\left(q^{-2d\delta\ell}+3q^{d-n}\right)\alpha\beta,

where the third transition uses Lemma A.12 and the fact that F=dF^{=d} is basis invariant by Lemma A.10, and finally the fourth transition uses Lemma A.13. Thus, the total contribution from the d>rd>r terms is

d=r+12|𝒯F=d,G=d|d=r+122qdδαβqrδαβ.\sum_{d=r+1}^{2\ell}\left|\langle\mathcal{T}^{\prime}F^{=d},G^{=d}\rangle\right|\leqslant\sum_{d=r+1}^{2\ell}2q^{-d\delta\ell}\sqrt{\alpha\beta}\leqslant q^{-r\delta\ell}\sqrt{\alpha\beta}.

Next, we bound the contribution from drd\leqslant r by bounding each term separately. Fix a drd\leqslant r. By Lemma A.15 and Holder’s inequality we have

|𝒯F=d,G=d|=|𝒯F=d,G|𝒯F=dtGt/(t1)β(t1)/tF=dtq500d2tβt1tεt1t,\left|\langle{\mathcal{T}^{\prime}F^{=d}},{G^{=d}}\rangle\right|=\left|\langle{\mathcal{T}^{\prime}F^{=d}},{G}\rangle\right|\leqslant\left\lVert\mathcal{T}^{\prime}F^{=d}\right\rVert_{t}\left\lVert G\right\rVert_{t/(t-1)}\leqslant\beta^{(t-1)/t}\left\lVert F^{=d}\right\rVert_{t}\leqslant q^{500d^{2}t}\beta^{\frac{t-1}{t}}\varepsilon^{\frac{t-1}{t}},

where in the last inequality we are using the fact that εα\varepsilon\geqslant\alpha and FF is (r,ε)(r,\varepsilon)-pseudo-random, so by Theorem A.7

F=dtq500d2tεt1t.\left\lVert F^{=d}\right\rVert_{t}\leqslant q^{500d^{2}t}\varepsilon^{\frac{t-1}{t}}.

Altogether, this shows

𝒯F,GqOt,r(1)βt1tεt1t+qrδαβ.\langle{\mathcal{T}^{\prime}F},{G}\rangle\leqslant q^{O_{t,r}(1)}\beta^{\frac{t-1}{t}}\varepsilon^{\frac{t-1}{t}}+q^{-r\delta\ell}\sqrt{\alpha\beta}.\qed

A.3 Reduction to the Bilinear Scheme

We are now ready to prove Lemma 2.3. As in the statement of Lemma 2.3, let FL2(𝖦𝗋𝖺𝗌𝗌q(n,2))F\in L_{2}({\sf Grass}_{q}(n,2\ell)) and GL2(𝖦𝗋𝖺𝗌𝗌q(n,2(1δ)))G\in L_{2}({\sf Grass}_{q}(n,2(1-\delta)\ell)) be Boolean functions, and suppose that FF is (r,ε)(r,\varepsilon)-pseudo-random. Define the Boolean functions FL2(𝔽qn×2)F^{\prime}\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right), G(L2(𝔽qn×2))G^{\prime}\in\left(L_{2}(\mathbb{F}_{q}^{n\times 2\ell})\right) by

F(x1,,x2)={F(span(x1,,x2))ifdim(span(x1,,x2)=2,0,otherwise,F^{\prime}(x_{1},\ldots,x_{2\ell})=\begin{cases}F(\operatorname{span}(x_{1},\ldots,x_{2\ell}))&\text{if}\ \dim(\operatorname{span}(x_{1},\ldots,x_{2\ell})=2\ell,\\ 0,&\text{otherwise},\end{cases}

and

G(x1,,x2(1δ))={G(span(x1,,x2(1δ)))ifdim(span(x1,,x2(1δ))=2(1δ),0,otherwise.G^{\prime}(x_{1},\ldots,x_{2(1-\delta)\ell})=\begin{cases}G(\operatorname{span}(x_{1},\ldots,x_{2(1-\delta)\ell}))&\text{if}\ \dim(\operatorname{span}(x_{1},\ldots,x_{2(1-\delta)\ell})=2(1-\delta)\ell,\\ 0,&\text{otherwise}.\end{cases}

We note that FF^{\prime} and GG^{\prime} are basis invariant functions. Next, we will prove that FF^{\prime} is (r,2ε)(r,2\varepsilon) pseudo-random, and towards this end we begin with the following lemma that simplifies the type of zoom-ins and zoom-out combinations we have to consider for FF^{\prime}.

Lemma A.17.

For any (U,V),(X,Y)(U,V),(X,Y) such that dim(U)+codim(X)<2\dim(U)+\operatorname{codim}(X)<2\ell and 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] is nonempty, there are rr^{\prime} linearly independent columns of VV, say v1,,vr𝔽qnv_{1},\ldots,v_{r^{\prime}}\in\mathbb{F}_{q}^{n} and a subset of linearly independent rows of XX, say X𝔽qs×nX^{\prime}\in\mathbb{F}_{q}^{s^{\prime}\times n}, such that rdim(U)r^{\prime}\leqslant\dim(U), scodim(X)s^{\prime}\leqslant\operatorname{codim}(X) and

F(U,V),(X,Y)222𝔼M𝔽qn×(2r)[F([v1,,vr,M])|XM=0],\left\lVert F^{\prime}_{(U,V),(X,Y)}\right\rVert_{2}^{2}\leqslant 2\cdot\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}\left([v_{1},\ldots,v_{r^{\prime}},M^{\prime}]\right)\;|\;X^{\prime}M^{\prime}=0],

where [v1,,vr,M]𝔽qn×2[v_{1},\ldots,v_{r^{\prime}},M^{\prime}]\in\mathbb{F}_{q}^{n\times 2\ell} is the matrix whose first rr columns are v1,,vrv_{1},\ldots,v_{r}, and remaining columns are MM^{\prime}.

Proof.

Let r=dim(U)r=\dim(U) and s=codim(X)s=\operatorname{codim}(X). First note that we can assume that the columns of UU and VV respectively are both nonzero and linearly independent. Indeed, otherwise say ui=0u_{i}=0, then either vi=0v_{i}=0, in which case the iith columns of UU and VV can be removed, or vi0v_{i}\neq 0 and 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] is an empty set. Otherwise, if, say, vi=0v_{i}=0, then either ui=0u_{i}=0 and again we can ignored the iith columns, or ui0u_{i}\neq 0 and 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] consists of matrices whose columns are not linearly independent. In this case F(U,V),(X,Y)F^{\prime}_{(U,V),(X,Y)} is identically 0 and the statement is trivially true. Similarly, if the columns of VV are not linearly independent, then 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] consists of matrices whose columns are not linearly independent, and again F(U,V),(X,Y)F^{\prime}_{(U,V),(X,Y)} is identically 0. Finally, if the columns of UU are not linearly independent, then either 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] is empty or there must be some ii such that both uiu_{i} and viv_{i} are linear combinations of the other columns in UU and VV respectively, with the same coefficients. In this case, we can remove the iith columns of UU and VV without changing 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)].

Now suppose that the columns of UU and VV are nonzero and linearly independent, and let A𝔽q2×2A\in\mathbb{F}_{q}^{2\ell\times 2\ell} be a full rank matrix such that AUi=eiAU_{i}=e_{i} for 1ir1\leqslant i\leqslant r. Let Y𝔽qn×(2r)Y^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r)} denote the last 2r2\ell-r columns of YA1YA^{-1}, and let Y′′Y^{\prime\prime} denote the first rr columns of YA1YA^{-1}. Since we assumed 𝖹𝗈𝗈𝗆[(U,V),(X,Y)]{\sf Zoom}[(U,V),(X,Y)] is nonempty, we must have X[v1,,vr]=Y′′X[v_{1},\ldots,v_{r}]=Y^{\prime\prime}. Then by the fact that FF^{\prime} is basis invariant we get that

F(U,V),(X,Y)22\displaystyle\left\lVert F^{\prime}_{(U,V),(X,Y)}\right\rVert_{2}^{2} =𝔼M𝔽qn×2[F(M)|MU=V,XM=Y]\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}[F^{\prime}(M)\;|\;MU=V,XM=Y]
=𝔼M𝔽qn×2[F(MA1)|MU=V,XM=Y]\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}[F^{\prime}(MA^{-1})\;|\;MU=V,XM=Y]
=𝔼M𝔽qn×2[F(M)|MAU=V,XM=Y]\displaystyle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell}}[F^{\prime}(M)\;|\;MAU=V,XM=Y^{\prime}]
=𝔼M𝔽qn×(2r)[F([v1,,vr,M])|XM=Y].\displaystyle=\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;XM^{\prime}=Y^{\prime}].

To complete the proof, we show to reduce to the case that YY^{\prime} is the zero matrix. First note that, using the same reasoning as we did for UU and VV, we can assume that the nonzero rows YY^{\prime} are linearly independent and the rows of XX are linearly independent

Suppose that y1,,ya𝔽q2ry^{\prime}_{1},\ldots,y^{\prime}_{a}\in\mathbb{F}_{q}^{2\ell-r} are the nonzero (and linearly independent) rows of YY^{\prime}, while the remaining rows are ya+1,,ys=0y^{\prime}_{a+1},\ldots,y^{\prime}_{s}=0. Let Y′′𝔽qa×(2r)Y^{\prime\prime}\in\mathbb{F}_{q}^{a\times(2\ell-r)} be the first a rows of YY^{\prime}, which are nonzero, let X𝔽qa×nX^{\prime}\in\mathbb{F}_{q}^{a\times n} denote the first aa rows of XX, and let X′′𝔽q(2a)×nX^{\prime\prime}\in\mathbb{F}_{q}^{(2\ell-a)\times n} denote rows a+1a+1 through ss of XX. For any Z=(z1,,za)𝔽qa×(2r)Z=(z_{1},\ldots,z_{a})\in\mathbb{F}_{q}^{a\times(2\ell-r)} with aa linearly independent rows let AZ𝔽q(2r)×(2r)A_{Z}\in\mathbb{F}_{q}^{(2\ell-r)\times(2\ell-r)} be the full rank matrix such that Y′′AZ=ZY^{\prime\prime}A_{Z}=Z. Then, for any linearly independent z1,,za𝔽q2rz_{1},\ldots,z_{a}\in\mathbb{F}_{q}^{2\ell-r},

𝔼M𝔽qn×(2r)[F([v1,,vr,M])|XM=Y]\displaystyle\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;XM^{\prime}=Y^{\prime}]
=𝔼M[F([v1,,vr,M])|XM=Y′′,X′′M=0]\displaystyle\qquad\qquad=\mathop{\mathbb{E}}_{M^{\prime}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;X^{\prime}M^{\prime}=Y^{\prime\prime},X^{\prime\prime}M^{\prime}=0]
=𝔼M[F([v1,,vr,MAZ1])|XM=Y′′,X′′M=0]\displaystyle\qquad\qquad=\mathop{\mathbb{E}}_{M^{\prime}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}A_{Z}^{-1}])\;|\;X^{\prime}M^{\prime}=Y^{\prime\prime},X^{\prime\prime}M^{\prime}=0]
=𝔼M[F([v1,,vr,M])|XM=Z,X′′M=0].\displaystyle\qquad\qquad=\mathop{\mathbb{E}}_{M^{\prime}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;X^{\prime}M^{\prime}=Z,X^{\prime\prime}M^{\prime}=0].

In the second transition, we used the fact that FF^{\prime} is basis invariant and multiplied its input by the matrix whose top left r×rr\times r minor is the identity, its bottom right (2r)×(2r)(2\ell-r)\times(2\ell-r) is AZ1A_{Z}^{-1}, and the rest of the entries are 0. Since the above holds for any ZZ with aa-linearly independent rows, letting EE denote the event that XMX^{\prime}M^{\prime} has aa linearly independent rows, it follows that

𝔼M𝔽qn×(2r)[F([v1,,vr,M])|XM=Y]=𝔼M[F([v1,,vr,M])|EX′′M=0],\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;XM^{\prime}=Y^{\prime}]=\mathop{\mathbb{E}}_{M^{\prime}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;E\land X^{\prime\prime}M^{\prime}=0],

and

𝔼M𝔽qn×(2r)[F([v1,,vr,M])|X′′M=0]\displaystyle\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;X^{\prime\prime}M^{\prime}=0] PrM𝔽qn×(2r)[E|X′′M=0]\displaystyle\geqslant\Pr_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[E\;|\;X^{\prime\prime}M^{\prime}=0]
𝔼M[F([v1,,vr,M])|EXM=Y].\displaystyle\cdot\mathop{\mathbb{E}}_{M^{\prime}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;E\land XM^{\prime}=Y^{\prime}].

Finally since Pr[E|X′′M=0]12\Pr[E\;|\;X^{\prime\prime}M^{\prime}=0]\geqslant\frac{1}{2} (as it is the probability of choosing a<2ra<2\ell-r linearly independent vectors in 𝔽q2r\mathbb{F}_{q}^{2\ell-r}), we have,

F(U,V),(X,Y)22\displaystyle\left\lVert F^{\prime}_{(U,V),(X,Y)}\right\rVert_{2}^{2} =𝔼M𝔽qn×(2r)[F([v1,,vr,M])|EXM=Y]\displaystyle=\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;E\land XM^{\prime}=Y^{\prime}]
2𝔼M[F([v1,,vr,M])|X′′M=0],\displaystyle\leqslant 2\mathop{\mathbb{E}}_{M^{\prime}}[F^{\prime}([v_{1},\ldots,v_{r},M^{\prime}])\;|\;X^{\prime\prime}M^{\prime}=0],

and the proof is concluded. ∎

As an immediate consequence of Lemma A.17

Lemma A.18.

If FF is (r,ε)(r,\varepsilon)-pseudo-random then FF^{\prime} is (r,2ε)(r,2\varepsilon)-pseudo-random.

Proof.

Fix any (U,V)(U,V) and (X,Y)(X,Y) such that dim(U)+codim(X)=r\dim(U)+\operatorname{codim}(X)=r. Using Lemma A.17, there are linearly independent v1,,vr𝔽qnv_{1},\ldots,v_{r^{\prime}}\in\mathbb{F}_{q}^{n} and X𝔽qs×nX^{\prime}\in\mathbb{F}_{q}^{s^{\prime}\times n} with linearly independent rows such that

F(U,V),(X,Y)22\displaystyle\left\lVert F^{\prime}_{(U,V),(X,Y)}\right\rVert_{2}^{2}
2𝔼M𝔽qn×(2r)[F([v1,,vr,M])|XM=0]\displaystyle\qquad\qquad\leqslant 2\cdot\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}\left([v_{1},\ldots,v_{r^{\prime}},M^{\prime}]\right)\;|\;X^{\prime}M^{\prime}=0]
2𝔼M𝔽qn×(2r)[F([v1,,vr,M])|XM=0,dim(im([v1,,vr,M]))=2]\displaystyle\qquad\qquad\leqslant 2\cdot\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F^{\prime}\left([v_{1},\ldots,v_{r^{\prime}},M^{\prime}]\right)\;|\;X^{\prime}M^{\prime}=0,\dim(\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}]))=2\ell]
=2𝔼M𝔽qn×(2r)[F(im([v1,,vr,M]))|XM=0,dim(im([v1,,vr,M]))=2],\displaystyle\qquad\qquad=2\cdot\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F\left(\operatorname{im}([v_{1},\ldots,v_{r^{\prime}},M^{\prime}])\right)\;|\;X^{\prime}M^{\prime}=0,\dim(\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}]))=2\ell],

where in the second transition we are using the fact that F(M)=0F^{\prime}(M)=0 for all MM such that dim(im(M))<2\dim(\operatorname{im}(M))<2\ell, and in the third transition we are using the definition of FF^{\prime}. We will bound the final term by using the pseudo-randomness of FF.

Choosing M𝔽qn×(2r)M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})} uniformly conditioned on XM=0X^{\prime}M^{\prime}=0, and dim(im([v1,,vr,M]))\dim(\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}])), we claim that im([v1,,vr,M])\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}]) is a uniformly random 22\ell-dimensional subspace in 𝖹𝗈𝗈𝗆[Q,QH]{\sf Zoom}[Q,Q\oplus H], where HH is the codimension ss subspace that is dual to the rows of XX^{\prime}. To see why, first note that it is clear im([v1,,vr,M])𝖹𝗈𝗈𝗆[Q,QH]\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}])\in{\sf Zoom}[Q,Q\oplus H]. Additionally, each L𝖹𝗈𝗈𝗆[Q,QH]L\in{\sf Zoom}[Q,Q\oplus H] has an equal number of M𝔽qn×(2r)M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})} such that

L=im([v1,,vr,M]),L=\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}]),

and therefore has an equal chance of being selected. It follows that choosing M𝔽qn×(2r)M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})} uniformly conditioned on XM=0X^{\prime}M^{\prime}=0, and dim(im([v1,,vr,M]))\dim(\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}])), im([v1,,vr,M])\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}]) is a uniformly random 22\ell-dimensional subspace in 𝖹𝗈𝗈𝗆[Q,QH]{\sf Zoom}[Q,Q\oplus H]. As a result,

F(U,V),(X,Y)22\displaystyle\left\lVert F^{\prime}_{(U,V),(X,Y)}\right\rVert_{2}^{2}
2𝔼M𝔽qn×(2r)[F(im([v1,,vr,M]))|XM=0,dim(im([v1,,vr,M]))=2]\displaystyle\qquad\qquad\leqslant 2\cdot\mathop{\mathbb{E}}_{M^{\prime}\in\mathbb{F}_{q}^{n\times(2\ell-r^{\prime})}}[F\left(\operatorname{im}([v_{1},\ldots,v_{r^{\prime}},M^{\prime}])\right)\;|\;X^{\prime}M^{\prime}=0,\dim(\operatorname{im}([v_{1},\ldots,v_{r},M^{\prime}]))=2\ell]
=2𝔼L𝖹𝗈𝗈𝗆[Q,QH][F(L)]\displaystyle\qquad\qquad=2\cdot\mathop{\mathbb{E}}_{L\in{\sf Zoom}[Q,Q\oplus H]}[F(L)]
2ε,\displaystyle\qquad\qquad\leqslant 2\varepsilon,

where in the last transition we use the fact that FF is (r,ε)(r,\varepsilon)-pseudo-random and dim(Q)+codim(QH)r+sr\dim(Q)+\operatorname{codim}(Q\oplus H)\leqslant r^{\prime}+s\leqslant r. ∎

Next, we note that the values of 𝒯F,G\langle\mathcal{T}F,G\rangle and 𝒯F,G\langle\mathcal{T}^{\prime}F,G\rangle are similar.

Lemma A.19.

We have

𝒯F,G2𝒯F,G.\langle\mathcal{T}F,G\rangle\leqslant 2\langle\mathcal{T}^{\prime}F^{\prime},G^{\prime}\rangle.
Proof.

We have,

𝒯F,G=𝔼M𝔽qn×2,A𝔽q2×2(1δ)[F(M)G(MA)|rank(A)=2(1δ)]\displaystyle\langle\mathcal{T}^{\prime}F^{\prime},G^{\prime}\rangle=\mathop{\mathbb{E}}_{M\in\mathbb{F}_{q}^{n\times 2\ell},A\in\mathbb{F}_{q}^{2\ell\times 2(1-\delta)\ell}}[F^{\prime}(M)\cdot G^{\prime}(MA)\;|\;\operatorname{rank}(A)=2(1-\delta)\ell]
PrM,A[rank(M)=2,rank(MA)=2(1δ)|rank(A)=2(1δ)]\displaystyle\quad\geqslant\Pr_{M,A}[\operatorname{rank}(M)=2\ell,\operatorname{rank}(MA)=2(1-\delta)\ell\;|\;\operatorname{rank}(A)=2(1-\delta)\ell]
𝔼M,A[F(M)G(MA)|rank(M)=2,rank(MA)=2(1δ),rank(A)=2(1δ)]\displaystyle\quad\cdot\mathop{\mathbb{E}}_{M,A}[F^{\prime}(M)\cdot G^{\prime}(MA)\;|\;\operatorname{rank}(M)=2\ell,\operatorname{rank}(MA)=2(1-\delta)\ell,\operatorname{rank}(A)=2(1-\delta)\ell]
=12𝔼M,A[F(M)G(MA)|rank(M)=2,rank(MA)=2(1δ),rank(A)=2(1δ)]\displaystyle\quad=\frac{1}{2}\mathop{\mathbb{E}}_{M,A}[F^{\prime}(M)\cdot G^{\prime}(MA)\;|\;\operatorname{rank}(M)=2\ell,\operatorname{rank}(MA)=2(1-\delta)\ell,\operatorname{rank}(A)=2(1-\delta)\ell]
=12𝔼M,A[F(im(M))G(im(MA))|rank(M)=2,rank(MA)=2(1δ),rank(A)=2(1δ)].\displaystyle\quad=\frac{1}{2}\mathop{\mathbb{E}}_{M,A}[F(\operatorname{im}(M))\cdot G(\operatorname{im}(MA))\;|\;\operatorname{rank}(M)=2\ell,\operatorname{rank}(MA)=2(1-\delta)\ell,\operatorname{rank}(A)=2(1-\delta)\ell].

To finish the proof, notice that in the conditional distribution (im(M),im(MA))(\operatorname{im}(M),\operatorname{im}(MA)) in the last term, im(M)\operatorname{im}(M) is a uniform L𝖦𝗋𝖺𝗌𝗌q(n,2)L\in{\sf Grass}_{q}(n,2\ell) and im(MA)\operatorname{im}(MA) is a uniform L𝖦𝗋𝖺𝗌𝗌q(n,2(1δ))L^{\prime}\in{\sf Grass}_{q}(n,2(1-\delta)\ell) such that LLL^{\prime}\subseteq L. Therefore,

𝒯F,G12𝔼L,L[F(L)G(L)|LL]=12𝒯F,G.\langle\mathcal{T}^{\prime}F^{\prime},G^{\prime}\rangle\geqslant\frac{1}{2}\mathop{\mathbb{E}}_{L,L^{\prime}}[F(L)\cdot G(L^{\prime})\;|\;L\supseteq L^{\prime}]=\frac{1}{2}\langle\mathcal{T}F,G\rangle.\qed

Lemma 2.3 follows by combining Lemma A.18 and Lemma A.19.

Proof of Lemma 2.3.

Suppose FL2(𝖦𝗋𝖺𝗌𝗌q(n,2)),GL2(𝖦𝗋𝖺𝗌𝗌q(n,2(1δ)))F\in L_{2}({\sf Grass}_{q}(n,2\ell)),G\in L_{2}({\sf Grass}_{q}(n,2(1-\delta)\ell)) have expectations α,β\alpha,\beta respectively, and suppose that FF is (r,ε)(r,\varepsilon) pseudo-random. Define the associated functions FL2(𝔽qn×2)F^{\prime}\in L_{2}\left(\mathbb{F}_{q}^{n\times 2\ell}\right) and GL2(𝔽qn×2(1δ))G^{\prime}\in L_{2}\left(\mathbb{F}_{q}^{n\times 2(1-\delta)\ell}\right) as above. It is clear that,

F22F22=α,G22G22=β.\left\lVert F^{\prime}\right\rVert_{2}^{2}\leqslant\left\lVert F\right\rVert_{2}^{2}=\alpha,\quad\left\lVert G^{\prime}\right\rVert_{2}^{2}\leqslant\left\lVert G\right\rVert_{2}^{2}=\beta.

Furthermore, by Lemma A.19, we have that

𝒯F,G2𝒯F,G.\langle\mathcal{T}F,G\rangle\leqslant 2\langle\mathcal{T}F^{\prime},G^{\prime}\rangle.

By Lemma A.18 FF^{\prime} is (r,2ε)(r,2\varepsilon)-pseudo-random, and applying Lemma A.16 we get that

𝒯F,G2𝒯F,GqOt,r(1)β(t1)/tε2t/(2t1)+qrδαβ.\langle\mathcal{T}F,G\rangle\leqslant 2\langle\mathcal{T}^{\prime}F^{\prime},G^{\prime}\rangle\leqslant q^{O_{t,r}(1)}\beta^{(t-1)/t}\varepsilon^{2t/(2t-1)}+q^{-r\delta\ell}\sqrt{\alpha\beta}.\qed

A.4 Proof of Lemma 2.4

We will show that if a set of \ell^{\prime}-dimensional subspaces 𝖦𝗋𝖺𝗌𝗌q(V,)\mathcal{L}^{\star}\subseteq{\sf Grass}_{q}(V^{\star},\ell^{\prime}) is pseudo-random, then it must “evenly cover” the space VV in the sense that there are very few points zVz\in V such that μz()\mu_{z}(\mathcal{L}^{\star}) significantly deviates from μ()\mu(\mathcal{L}^{\star}). We will require the following result from [EKL24].

Theorem A.20.

[EKL24, Theorem 1.12] If FL2(𝔽qn×)F^{\prime}\in L_{2}\left(\mathbb{F}_{q}^{n\times\ell^{\prime}}\right) is a Boolean function which is (1,ε)(1,\varepsilon)-global, then for all powers of t4t\geqslant 4 that are powers of 22 it holds that

F=122q460tF222tε.\left\lVert F^{\prime=1}\right\rVert_{2}^{2}\leqslant q^{460t}\left\lVert F\right\rVert_{2}^{2-\frac{2}{t}}\varepsilon.
Lemma A.21.

Let 𝖦𝗋𝖺𝗌𝗌q(V,)\mathcal{L}^{\star}\subseteq{\sf Grass}_{q}(V^{\star},\ell^{\prime}) have μ()=ηqC\mu(\mathcal{L}^{\star})=\eta\geqslant q^{-C\ell^{\prime}} for some large constant CC, and set Z={zV||μz()η|η10}Z=\{z\in V^{\star}\;|\;|\mu_{z}(\mathcal{L}^{\star})-\eta|\leqslant\frac{\eta}{10}\}. If \mathcal{L}^{\star} is (1,qcη)(1,q^{c\ell^{\prime}}\eta)-pseudo-random for some 0<c<10<c<1, then

|Z|(1q2)|V|.|Z|\geqslant\left(1-q^{\frac{\ell^{\prime}}{2}}\right)|V^{\star}|.
Proof.

Let dim(V)=n\dim(V^{\star})=n, let FL2(𝖦𝗋𝖺𝗌𝗌q(V,))F\in L_{2}({\sf Grass}_{q}(V^{\star},\ell^{\prime})) be the indicator function for \mathcal{L}^{\star}, and let FL2(𝔽qn×)F^{\prime}\in L_{2}\left(\mathbb{F}_{q}^{n\times\ell^{\prime}}\right) be the associated function given by

F(x1,,x)={F(span(x1,,x))ifdim(span(x1,,x2))=,0otherwise.F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}})=\begin{cases}F(\operatorname{span}(x_{1},\ldots,x_{\ell^{\prime}}))&\text{if}\ \dim(\operatorname{span}(x_{1},\ldots,x_{2\ell}))=\ell^{\prime},\\ 0&\text{otherwise}.\end{cases}

By Lemma A.18, FF^{\prime} is (1,2qcη)(1,2q^{c\ell^{\prime}}\eta)-pseudo-random. For any point xVx\in V^{\star}, we have

μz()=𝔼x1,,x1V[F(x1,,x1,z)|dim(span(x1,,x1,z))=],\mu_{z}(\mathcal{L}^{\star})=\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}\in V^{\star}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\;|\;\dim(\operatorname{span}(x_{1},\ldots,x_{\ell^{\prime}-1},z))=\ell^{\prime}\right],

so it follows that

|μz()𝔼x1,,x1[F(x1,,x1,z)]|qqnand|μ()F22|qqn.\left|\mu_{z}(\mathcal{L}^{\star})-\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\right]\right|\leqslant\frac{q^{\ell^{\prime}}}{q^{n}}\quad\text{and}\quad\left|\mu(\mathcal{L}^{\star})-\left\lVert F^{\prime}\right\rVert_{2}^{2}\right|\leqslant\frac{q^{\ell^{\prime}}}{q^{n}}.

Thus

|𝔼x1,,x1[F(x1,,x1,z)]F22||μz()μ()|qqn.\left|\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\right]-\left\lVert F^{\prime}\right\rVert_{2}^{2}\right|\geqslant\left|\mu_{z}(\mathcal{L}^{\star})-\mu(\mathcal{L}^{\star})\right|-\frac{q^{\ell^{\prime}}}{q^{n}}. (19)

We will now relate this quantity to the level one weight of FF^{\prime} and apply Lemma A.7 to bound the level one weight of FF^{\prime}. Note that

𝔼x1,,x1[F(x1,,x1,z)]\displaystyle\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\right] =S=(s1,,s)VF^(S)𝔼x1,,x1[χS(x1,,x1,z)]\displaystyle=\sum_{S=(s_{1},\ldots,s_{\ell^{\prime}})\in V^{\star}}\widehat{F^{\prime}}(S)\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}[\chi_{S}(x_{1},\ldots,x_{\ell^{\prime}-1},z)]
=S=(s1,,s)VF^(S)χs(z)i=11𝔼xi[χsi(xi)].\displaystyle=\sum_{S=(s_{1},\ldots,s_{\ell^{\prime}})\in V^{\star}}\widehat{F^{\prime}}(S)\chi_{s_{\ell^{\prime}}}(z)\prod_{i=1}^{\ell^{\prime}-1}\mathop{\mathbb{E}}_{x_{i}}\left[\chi_{s_{i}}(x_{i})\right].

Now note that 𝔼xi[χsi(xi)]=0\mathop{\mathbb{E}}_{x_{i}}\left[\chi_{s_{i}}(x_{i})\right]=0 if sis_{i} is not the zero vector, and 𝔼xi[χsi(xi)]=1\mathop{\mathbb{E}}_{x_{i}}\left[\chi_{s_{i}}(x_{i})\right]=1 if sis_{i} is the zero vector. Thus,

𝔼x1,,x1[F(x1,,x1,z)]\displaystyle\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\right] =F^(0,,0)+aV,a0F^(0,,0,a)χa(z),\displaystyle=\widehat{F^{\prime}}(0,\ldots,0)+\sum_{a\in V^{\star},a\neq 0}\widehat{F}(0,\ldots,0,a)\chi_{a}(z),

and using the fact that F^(0,,0)=𝔼[F]=F22\widehat{F^{\prime}}(0,\ldots,0)=\mathop{\mathbb{E}}[F^{\prime}]=\left\lVert F^{\prime}\right\rVert_{2}^{2},

(𝔼x1,,x1[F(x1,,x1,z)]F22)2=(aV,a0F^(0,,0,a)χa(z))2.\left(\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\right]-\left\lVert F^{\prime}\right\rVert_{2}^{2}\right)^{2}=\left(\sum_{a\in V^{\star},a\neq 0}\widehat{F}(0,\ldots,0,a)\chi_{a}(z)\right)^{2}.

Therefore, we get by (19) that

𝔼zV[(μz()η)2]\displaystyle\mathop{\mathbb{E}}_{z\in V^{\star}}[(\mu_{z}(\mathcal{L}^{\star})-\eta)^{2}] 𝔼zV[|𝔼x1,,x1[F(x1,,x1,z)]F22|2]+5qqn\displaystyle\leqslant\mathop{\mathbb{E}}_{z\in V^{\star}}\left[\left|\mathop{\mathbb{E}}_{x_{1},\ldots,x_{\ell^{\prime}-1}}\left[F^{\prime}(x_{1},\ldots,x_{\ell^{\prime}-1},z)\right]-\left\lVert F^{\prime}\right\rVert_{2}^{2}\right|^{2}\right]+5\frac{q^{\ell^{\prime}}}{q^{n}}
𝔼z[|aV,a0F^(0,,0,a)χa(z)|2]+5qqn\displaystyle\leqslant\mathop{\mathbb{E}}_{z}\left[\left|\sum_{a\in V^{\star},a\neq 0}\widehat{F}(0,\ldots,0,a)\chi_{a}(z)\right|^{2}\right]+5\frac{q^{\ell^{\prime}}}{q^{n}}
=aV,a0|F^(0,,0,a)|2+5qqn.\displaystyle=\sum_{a\in V^{\star},a\neq 0}|\widehat{F}(0,\ldots,0,a)|^{2}+5\frac{q^{\ell^{\prime}}}{q^{n}}.

Next since FF^{\prime} is basis invariant, using Lemma A.8, we have that for all α1,,α𝔽q\alpha_{1},\ldots,\alpha_{\ell^{\prime}}\in\mathbb{F}_{q} that are not all zero,

F^(α1a,,αa)=F^(0,,0,a).\displaystyle\widehat{F^{\prime}}\left(\alpha_{1}a,\ldots,\alpha_{\ell^{\prime}}a\right)=\widehat{F^{\prime}}\left(0,\ldots,0,a\right).

It follows that

aV,a0|F^(0,,0,a)|2=1q1rank(S)=1|F^(S)|2=F=122q1.\sum_{a\in V^{\star},a\neq 0}|\widehat{F^{\prime}}(0,\ldots,0,a)|^{2}=\frac{1}{q^{\ell^{\prime}}-1}\sum_{\operatorname{rank}(S)=1}|\widehat{F^{\prime}}(S)|^{2}=\frac{\left\lVert F^{\prime=1}\right\rVert_{2}^{2}}{q^{\ell^{\prime}}-1}.

Using Theorem A.20 with d=1d=1, along the fact that FF^{\prime} is (1,qcη)(1,q^{c\ell^{\prime}}\eta)-pseudo-random, we get

𝔼zV[(μz()η)2]F=122q1+5qqnq460tqcη22tq1,\mathop{\mathbb{E}}_{z\in V^{\star}}[(\mu_{z}(\mathcal{L}^{\star})-\eta)^{2}]\leqslant\frac{\left\lVert F^{\prime=1}\right\rVert_{2}^{2}}{q^{\ell^{\prime}}-1}+5\frac{q^{\ell^{\prime}}}{q^{n}}\leqslant\frac{q^{460t}q^{c\ell^{\prime}}\eta^{2-\frac{2}{t}}}{q^{\ell^{\prime}}-1},

for any t4t\geqslant 4 that is a power of 22. By Markov’s inequality it follows that

|Z¯|qnη2100100q460tqcη2tqη2100q460tqcq(1c2Ct)η2q2100η2,\frac{|\overline{Z}|}{q^{n}}\cdot\frac{\eta^{2}}{100}\leqslant\frac{100q^{460t}q^{c\ell^{\prime}}}{\eta^{\frac{2}{t}}q^{\ell^{\prime}}}\eta^{2}\leqslant\frac{100q^{460t}q^{c\ell^{\prime}}}{q^{\left(1-c-\frac{2C}{t}\right)\ell^{\prime}}}\eta^{2}\leqslant\frac{q^{-\frac{\ell^{\prime}}{2}}}{100}\eta^{2},

where we take tt to be a power of 22 large enough so that 1c+2Ct121-c+\frac{2C}{t}\geqslant\frac{1}{2}. Dividing by η2\eta^{2} finishes the proof. ∎

Appendix B Proof of Theorem 5.2

In order to prove Theorem 5.2 we will find the subspaces, QQ, one at at time by using Theorem 5.1. We let 𝒬\mathcal{Q} denote the set of all QQ’s collected thus far. Each time a new subspace QQ is added to 𝒬\mathcal{Q}, we randomize the assignment T1[L]T_{1}[L] for all 22\ell-dimensional LQL\supset Q. At a high level, the effect of this randomization is that there is only a little agreement between any linear function and the assignments on subspaces containing QQ, thus these entries are essentially “deleted”.

More formally, we construct the set 𝒬\mathcal{Q} of subspaces as follows. Initially set T~1=T1\tilde{T}_{1}=T_{1}, 𝒬=\mathcal{Q}=\emptyset, and X=X=\emptyset. Recall that initially T~1\tilde{T}_{1} and T2T_{2} are ε\varepsilon-consistent for ε2q2(11000δ)\varepsilon\geqslant 2q^{-2\ell(1-1000\delta)}. While T~1\tilde{T}_{1} and T2T_{2} are at least ε/2\varepsilon/2-consistent, do the following.

  1. 1.

    Let QWQ\subset W be subspaces guaranteed by Theorem 5.1. That is, QQ and WW satisfy dim(Q)+codim(W)=r\dim(Q)+\operatorname{codim}(W)=r and there exists linear gQ,W:W𝔽qg_{Q,W}:\xrightarrow[]{}W\xrightarrow[]{}\mathbb{F}_{q} such that

    PrL𝖦𝗋𝖺𝗌𝗌(n,2)[gQ,W|L=T~1[L]|QLW]ε.\Pr_{L\in{\sf Grass}(n,2\ell)}[g_{Q,W}|_{L}=\tilde{T}_{1}[L]\;|\;Q\subseteq L\subseteq W]\geqslant\varepsilon^{\prime}.
  2. 2.

    Set 𝒬𝒬{Q}\mathcal{Q}\xleftarrow[]{}\mathcal{Q}\cup\{Q\}.

  3. 3.

    Set XX{L|QLW}X\xleftarrow[]{}X\cup\{L\;|\;Q\subseteq L\subseteq W\}.

  4. 4.

    For each LXL\in X independently, choose T~1[L]\tilde{T}_{1}[L] uniformly among all linear functions on LL.

We have the following claim regarding the re-assignment phase.

Claim B.1.

With probability 1eΩ(qn)1-e^{-\Omega(q^{\ell n})} over the random assignment on XX, for every Q,WQ,W such that QWQ\subset W and dim(Q)+codim(W)=r\dim(Q)+\operatorname{codim}(W)=r at least one of the following holds:

  1. 1.

    Less than q2q^{-2\ell}-fraction of L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W] are in XX,

  2. 2.

    For every linear function gQ,W:W𝔽qg_{Q,W}:W\xrightarrow[]{}\mathbb{F}_{q},

    PrLX,L𝖹𝗈𝗈𝗆[Q,W][gQ,WT~1[L]]q12\Pr_{L\in X,L\in{\sf Zoom}[Q,W]}[g_{Q,W}\equiv\tilde{T}_{1}[L]]\leqslant q^{1-2\ell}
Proof.

Note that the first item has nothing to do with the random assignment over XX, so we need only show that if the first item is false then the second item must be true. Suppose that the first item does not hold and that XX contains at least q2q^{-2\ell}-fraction of L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W].

Fix Q,W,gQ,WQ,W,g_{Q,W} with the parameters above and let A=𝖹𝗈𝗈𝗆[Q,W]XA={\sf Zoom}[Q,W]\cap X, and suppose that the first item does not hold. We will show that in this case, the second item holds.

For each LAL\in A, let ZLZ_{L} denote the indicator variable that takes value 11 if gQ,W|L=T~1[L]g_{Q,W}|_{L}=\tilde{T}_{1}[L] and 0 otherwise. The expectation of ZLZ_{L} is q2q^{-2\ell}, and by a Chernoff bound, the desired probability is bounded by

PrLA[1|A|LAZLq12]eq2+2|A|/6.\Pr_{L\in A}\left[\frac{1}{|A|}\sum_{L\in A}Z_{L}\geqslant q^{1-2\ell}\right]\leqslant e^{-q^{-2\ell+2}|A|/6}.

By assumption, |A|q2|{L𝖹𝗈𝗈𝗆[Q,W],dim(L)=2}|q2q(2r)(nr2)|A|\geqslant q^{-2\ell}\left|\{L\in{\sf Zoom}[Q,W],\dim(L)=2\ell\}\right|\geqslant q^{-2\ell}q^{(2\ell-r)(n-r-2\ell)}. Thus, using a union bound over all Q,W,gQ,WQ,W,g_{Q,W}, the probability that there exist a bad triple is at most,

(r+1)qnrqneq4+2q(2r)(nr2)/6eΩ(qn).(r+1)q^{nr}q^{n}e^{-q^{-4\ell+2}q^{(2\ell-r)(n-r-2\ell)}/6}\leqslant e^{-\Omega(q^{\ell n})}.\qed

We now analyze the process. Note that using Chernoff’s bound, with probability 1eΩ(qn)1-e^{-\Omega(q^{\ell n})} over the randomization step, the probability PrL𝖦𝗋𝖺𝗌𝗌(n,2)[gQ,W|L=T~1[L]|QLW]\Pr_{L\in{\sf Grass}(n,2\ell)}[g_{Q,W}|_{L}=\tilde{T}_{1}[L]\;|\;Q\subseteq L\subseteq W] drops from at least ε\varepsilon^{\prime} to at most q12q^{1-2\ell}. In that case, the measure of XX increases by at least

(εq12)qrnqO(rn).(\varepsilon^{\prime}-q^{1-2\ell})q^{-rn}\geqslant q^{-O(rn)}.

Doing a union bound over the steps, it follows that with probability 1eΩ(qn)qO(rn)=1o(1)1-e^{-\Omega(q^{\ell n})}q^{O(rn)}=1-o(1) the process terminates within qO(rn)q^{O(rn)} steps.

Note that it is possible that the same subspace QQ is added multiple times (with different zoom-outs) in the process above, so we clarify that 𝒬\mathcal{Q} is considered as a set without repeats. Also note that with probability 1o(1)1-o(1), for each Q𝒬Q\in\mathcal{Q}, WW and gQ,Wg_{Q,W} found in the process it holds that

PrL𝖦𝗋𝖺𝗌𝗌(n,2)[gQ,W|L=T1[L]|QLW]12ε\Pr_{L\in{\sf Grass}(n,2\ell)}[g_{Q,W}|_{L}=T_{1}[L]\;|\;Q\subseteq L\subseteq W]\geqslant\frac{1}{2}\varepsilon^{\prime} (20)

(the point being is that the agreement now is compared to the original T1T_{1} and not to T~1\tilde{T}_{1}). Indeed, considering the step Q,WQ,W and gQ,Wg_{Q,W} were found, gQ,Wg_{Q,W} had agreement at least ε\varepsilon^{\prime} with T~1\tilde{T}_{1} on 𝖹𝗈𝗈𝗆[Q,W]{\sf Zoom}[Q,W] at that point, and by Claim B.1 with probability 1eΩ(qn)1-e^{-\Omega(q^{\ell n})} at most q12ε/2q^{1-2\ell}\leqslant\varepsilon^{\prime}/2 of that agreement came from LXL\in X. Thus, by union bound over all of the steps, with probability 1qO(rn)qΩ(n)=1o(1)1-q^{O(rn)}q^{-\Omega(\ell n)}=1-o(1) it follows that (20) holds for every Q,WQ,W and gQ,Wg_{Q,W} found throughout the process.

The following claim shows that at the end of the process the number of QQ’s found in the process is large, thereby finishing the proof of Theorem 5.2.

Claim B.2.

There exists some 0r1r0\leqslant r_{1}\leqslant r such that 𝒬\mathcal{Q} contains at least a q52q^{-5\ell^{2}}-fraction of all r1r_{1}-dimensional subspaces.

Proof.

At the end of the process, the consistency has dropped by at least ε/2\varepsilon^{\prime}/2, so the probability over edges (L,R)(L,R) that LL was reassigned must be at least ε/2\varepsilon^{\prime}/2. For each 0r1r0\leqslant r_{1}\leqslant r, let Nr1N_{r_{1}} be the number of QQ of dimension r1r_{1} in 𝒬\mathcal{Q}.

For each QQ of dimension r1r_{1}, the fraction of 22\ell-dimensional LL’s that are reassigned due to QQ being added to 𝒬\mathcal{Q} is at most the fraction of 22\ell-dimensional subspaces that contain QQ. This is,

[n2r1]q[n2]qqn(2r1)q2(n2)=q42r1n.\frac{\begin{bmatrix}{n}\\ {2\ell-r_{1}}\end{bmatrix}_{q}}{\begin{bmatrix}{n}\\ {2\ell}\end{bmatrix}_{q}}\leqslant\frac{q^{n(2\ell-r_{1})}}{q^{2\ell(n-2\ell)}}=q^{4\ell^{2}-r_{1}n}.

It follows that there must be some r1r_{1} such that

Nr1q42r1nεr+1.N_{r_{1}}q^{4\ell^{2}-r_{1}n}\geqslant\frac{\varepsilon}{r+1}.

Rearranging this inequality, we get that

Nr1ε(r+1)q42qr1nq52[nr1]q.N_{r_{1}}\geqslant\frac{\varepsilon}{(r+1)q^{4\ell^{2}}}q^{r_{1}}n\geqslant q^{-5\ell^{2}}\begin{bmatrix}{n}\\ {r_{1}}\end{bmatrix}_{q}.

Thus there exists an r1r_{1} such that 𝒬\mathcal{Q} contains at least a q52q^{-5\ell^{2}}-fraction of all r1r_{1}-dimensional subspaces. ∎

Appendix C The Covering Property

Fix a question UU to the first prover. Recall that we set

k=q2(1+c)andβ=q2(1+2c/3),k=q^{2(1+c)\ell}\quad\text{and}\quad\beta=q^{-2(1+2c/3)\ell},

where 0<c<10<c<1 is some small constant close to 0 and set η=q100100\eta=q^{-100\ell^{100}}, and recall that the distributions 𝒟\mathcal{D} and 𝒟\mathcal{D}^{\prime} are defined as follows.

𝒟:\mathcal{D}:

  • Choose x1,,x2𝔽qUx_{1},\ldots,x_{2\ell}\in\mathbb{F}_{q}^{U} uniformly.

  • Output the list (x1,,x2)(x_{1},\ldots,x_{2\ell}).

𝒟:\mathcal{D}^{\prime}:

  • Choose VUV\subseteq U according to the Outer PCP.

  • Choose x1,,x2𝔽qVx^{\prime}_{1},\ldots,x^{\prime}_{2\ell}\in\mathbb{F}_{q}^{V} uniformly, and lift these vectors to 𝔽qU\mathbb{F}_{q}^{U} by inserting 0’s into the missing coordinates.

  • Choose w1,,w2HUw_{1},\ldots,w_{2\ell}\in H_{U} uniformly, and set xi=xi+wix_{i}=x^{\prime}_{i}+w_{i} for 1i21\leqslant i\leqslant 2\ell.

  • Output the list (x1,,x2)(x_{1},\ldots,x_{2\ell}).

We also restate Lemma 5.4 below as a reminder. See 5.4

C.1 Proof of Lemma 5.4

For x1,,x2𝔽q|U|x_{1},\ldots,x_{2\ell}\in\mathbb{F}_{q}^{|U|}, let us view x1,,x2x_{1},\ldots,x_{2\ell} as the rows of a 2×3k2\ell\times 3k matrix, and split the columns of this matrix into kk blocks - each consisting of 33 consecutive columns. Then let s(x1,,x2)s(x_{1},\ldots,x_{2\ell}) be the number of blocks where at least two of the columns are equal, and set p=3q22q4p=3q^{-2\ell}-2q^{-4\ell}. The idea is that s(x1,,x2)s(x_{1},\ldots,x_{2\ell}) should correspond to the number of equations where we drop variables in the Outer PCP, while pp is the probability that a fixed block has at least two columns equal to each other. Also let s(x1,,x2)s^{\prime}(x_{1},\ldots,x_{2\ell}) be the number of blocks where all 33 columns are equal, and let p=q4p^{\prime}=q^{-4\ell}, be the probability that a fixed block has all three columns equal.

We define the set EE as follows. Set

E1={(x1,,x2)(𝔽qU)2||s(x1,,x2)pk|>50pklog(1/η)},E_{1}=\left\{(x_{1},\ldots,x_{2\ell})\in\left(\mathbb{F}_{q}^{U}\right)^{2\ell}\;|\;\left|s(x_{1},\ldots,x_{2\ell})-pk\right|>50\sqrt{pk\log(1/\eta)}\right\}, (21)
E2={(x1,,x2)(𝔽qU)2|s(x1,,x2)>100},E_{2}=\left\{(x_{1},\ldots,x_{2\ell})\in\left(\mathbb{F}_{q}^{U}\right)^{2\ell}\;|\;s^{\prime}(x_{1},\ldots,x_{2\ell})>\ell^{100}\right\}, (22)

and E=E1E2E=E_{1}\cup E_{2}.

By a Chernoff bound,

𝒟(E1)=Pr𝒟[|s(x1,,x2)pk|>50pklog(1/η)]η50,\mathcal{D}(E_{1})=\Pr_{\mathcal{D}}\left[|s(x_{1},\ldots,x_{2\ell})-pk|>50\sqrt{pk\log(1/\eta)}\right]\leqslant\eta^{50},

where recall that η=q100100\eta=q^{-100\ell^{100}}. Also, by our setting of βk\beta k, we have βk=q2c/3\beta k=q^{2c\ell/3}, while pk=O(q2c)pk=O(q^{2c\ell}), so the same Chernoff bound holds for 𝒟\mathcal{D}^{\prime},

𝒟(E1)=Pr𝒟[|s(x1,,x2)pk|>50pklog(1/η)]η50,\mathcal{D}^{\prime}(E_{1})=\Pr_{\mathcal{D}^{\prime}}\left[|s(x_{1},\ldots,x_{2\ell})-pk|>50\sqrt{pk\log(1/\eta)}\right]\leqslant\eta^{50},

so 𝒟(E1)η45\mathcal{D}^{\prime}(E_{1})\leqslant\eta^{45}. Indeed, the actual expectation of s(x1,,x2)s(x_{1},\ldots,x_{2\ell}) under 𝒟\mathcal{D}^{\prime} is (1β)pk+βk(1-\beta)pk+\beta k and this differs from pkpk by only O(βk)=O(pk)O(\beta k)=O(pk).

For the measure of E2E_{2} we have,

𝒟(E2)=PrD[s(x1,,x2)>100](k100)p100(kp)100η100,\mathcal{D}(E_{2})=\Pr_{D}[s^{\prime}(x_{1},\ldots,x_{2\ell})>\ell^{100}]\leqslant\binom{k}{\ell^{100}}p^{\prime\ell^{100}}\leqslant(kp^{\prime})^{\ell^{100}}\leqslant\eta^{100},

where in the middle term, the first factor is the number of ways to choose 100\ell^{100} blocks and the second factor is the probability that all of these blocks have all three columns equal. Similarly,

𝒟(E2)=PrD[s(x1,,x2)>100](k100)((1β)p+β1q2)100k100(q4)100η100.\mathcal{D}^{\prime}(E_{2})=\Pr_{D^{\prime}}[s^{\prime}(x_{1},\ldots,x_{2\ell})>\ell^{100}]\leqslant\binom{k}{\ell^{100}}\left((1-\beta)p^{\prime}+\beta\frac{1}{q^{2\ell}}\right)^{\ell^{100}}\leqslant k^{\ell^{100}}\left(q^{-4\ell}\right)^{\ell^{100}}\leqslant\eta^{100}.

Putting everything together, we get that

𝒟(E)𝒟(E1)+𝒟(E2)η40and𝒟(E)𝒟(E1)+𝒟(E2)η40.\mathcal{D}(E)\leqslant\mathcal{D}(E_{1})+\mathcal{D}(E_{2})\leqslant\eta^{40}\quad\text{and}\quad\mathcal{D}^{\prime}(E)\leqslant\mathcal{D}^{\prime}(E_{1})+\mathcal{D}^{\prime}(E_{2})\leqslant\eta^{40}. (23)

We next show that the probability measures 𝒟\mathcal{D} and 𝒟\mathcal{D}^{\prime} assign roughly the same measure to each xEx\not\in E. Fix (x1,,x2)E(x_{1},\ldots,x_{2\ell})\notin E. It is clear that D(x1,,x2)=q(2)3kD(x_{1},\ldots,x_{2\ell})=q^{-(2\ell)3k}, where we use |U|=3k|U|=3k. Let s=s(x1,,x2)s=s(x_{1},\ldots,x_{2\ell}) and s=s(x1,,x2)s^{\prime}=s^{\prime}(x_{1},\ldots,x_{2\ell}). Then,

D(x1,,x2)\displaystyle D^{\prime}(x_{1},\ldots,x_{2\ell}) =((1β)q32)kss((1β)q32+β3q2)s((1β)q32+βq22)s\displaystyle=((1-\beta)q^{-3\cdot 2\ell})^{k-s-s^{\prime}}\left((1-\beta)q^{-3\cdot 2\ell}+\frac{\beta}{3}q^{-2\cdot\ell}\right)^{s}\left((1-\beta)q^{-3\cdot 2\ell}+\beta q^{-2\cdot 2\ell}\right)^{s^{\prime}}
=q23k(1β)kss(1β+β3q2)s(1β+βq2)s\displaystyle=q^{-2\ell\cdot 3k}(1-\beta)^{k-s-s^{\prime}}\left(1-\beta+\frac{\beta}{3}q^{2\ell}\right)^{s}(1-\beta+\beta q^{2\ell})^{s^{\prime}} (24)

In the first equality, the first term is the probability of choosing the blocks that have three all distinct columns. Then, (1β)(1-\beta) is the probability that no variables are dropped, and q3(2)q^{-3(2\ell)} is the probability of choosing those three particular xix_{i}’s in that block. The second term is the probability of choosing the blocks that have exactly two equal columns. Then, (1β)q3(2)(1-\beta)q^{-3(2\ell)} is the probability of having no variables dropped and choosing the three xix_{i}’s, and β3q2(2)\frac{\beta}{3}q^{-2(2\ell)} is the probability of first having the variable dropped in the column that is not equal to the other two, and then choosing the correct values for the remaining two column values.

We will first show

D(x1,,x2)D(x1,,x2)11.1.\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})}\geqslant\frac{1}{1.1}.

Using Equation (C.1),

D(x1,,x2)D(x1,,x2)\displaystyle\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})} =(1β)kss(1β+β3q2)s(1β+βq2)s\displaystyle=(1-\beta)^{k-s-s^{\prime}}\left(1-\beta+\frac{\beta}{3}q^{2\ell}\right)^{s}(1-\beta+\beta q^{2\ell})^{s^{\prime}}
(1β)ks(1β+β3q2)s\displaystyle\geqslant(1-\beta)^{k-s}\left(1-\beta+\frac{\beta}{3}q^{2\ell}\right)^{s}
=(1β)ks(1+β(q231))s\displaystyle=(1-\beta)^{k-s}\left(1+\beta\left(\frac{q^{2\ell}}{3}-1\right)\right)^{s}
exp(β(ks)β2(ks)+(q231)βs(q231)2β2s)\displaystyle\geqslant\exp\left(-\beta(k-s)-\beta^{2}(k-s)+\left(\frac{q^{2\ell}}{3}-1\right)\beta s-\left(\frac{q^{2\ell}}{3}-1\right)^{2}\beta^{2}s\right)
=exp(βkβ2(ks)+q23βs(q231)2β2s)\displaystyle=\exp\left(-\beta k-\beta^{2}(k-s)+\frac{q^{2\ell}}{3}\beta s-\left(\frac{q^{2\ell}}{3}-1\right)^{2}\beta^{2}s\right)
exp(βk+q23βs(q231)2β2sq2),\displaystyle\geqslant\exp\left(-\beta k+\frac{q^{2\ell}}{3}\beta s-\left(\frac{q^{2\ell}}{3}-1\right)^{2}\beta^{2}s-q^{-2\ell}\right),

where in the fourth transition we use the bound 1+zezz21+z\geqslant e^{z-z^{2}}, which holds for all real numbers zz such that |z||z| is sufficiently small. For our uses, z=βz=\beta and z=β(q231)z=\beta\left(\frac{q^{2\ell}}{3}-1\right) are q2(1+2c/3)q^{-2(1+2c/3)\ell} and O(q2c/3)O(q^{-2c\ell/3}), and both are sufficiently small. In the last transition we use the fact that β2(ks)β2kq2-\beta^{2}(k-s)\geqslant-\beta^{2}k\geqslant-q^{-2\ell}.

Now write s=pkvs=pk-v and let us analyze the first two terms in the last line. Plugging this in and using the definition of pp we get that

βk+q23βs=βk+q23β(pkv)=βk(1q23p)q23βv=23βkq2q23βv.-\beta k+\frac{q^{2\ell}}{3}\beta s=-\beta k+\frac{q^{2\ell}}{3}\beta(pk-v)=-\beta k\left(1-\frac{q^{2\ell}}{3}p\right)-\frac{q^{2\ell}}{3}\beta v=-\frac{2}{3}\beta kq^{-2\ell}-\frac{q^{2\ell}}{3}\beta v. (25)

Plugging this back into the above,

D(x1,,x2)D(x1,,x2)\displaystyle\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})} exp((q231)2β2pk)\displaystyle\geqslant\exp\left(-\left(\frac{q^{2\ell}}{3}-1\right)^{2}\beta^{2}pk\right)
exp(q23βv+(q231)2β2v)exp(23βkq2).\displaystyle\cdot\exp\left(-\frac{q^{2\ell}}{3}\beta v+\left(\frac{q^{2\ell}}{3}-1\right)^{2}\beta^{2}v\right)\cdot\exp\left(-\frac{2}{3}\beta kq^{-2\ell}\right).

Plugging in our values for β,k\beta,k and pp, the first term on the right hand side above is exp(Θ(q2c/3))\exp\left(-\Theta(q^{-2c\ell/3})\right). Using v50pklog(1/η)q(c+o(1))v\leqslant 50\sqrt{pk\log(1/\eta)}\leqslant q^{(c+o(1))\ell}, the second term on the right hand side is at least exp(Ω(q(c/3+o(1))))\exp\left(-\Omega(q^{(-c/3+o(1))\ell})\right). Finally, the last term is at least exp(q2)\exp(q^{-2\ell}). Overall, we get that for large enough \ell

D(x1,,x2)D(x1,,x2)exp(O(q(c/3o(1))))11.1.\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})}\geqslant\exp\left(-O(q^{-(c/3-o(1))\ell})\right)\geqslant\frac{1}{1.1}.

For the other direction, we show

D(x1,,x2)D(x1,,x2)10.9,\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})}\leqslant\frac{1}{0.9},

in nearly the same fashion. First note that

(1β+βq21β)s(1β+βq21β)100=(1+β1βq2)1001+o(1).\left(\frac{1-\beta+\beta q^{2\ell}}{1-\beta}\right)^{s^{\prime}}\leqslant\left(\frac{1-\beta+\beta q^{2\ell}}{1-\beta}\right)^{\ell^{100}}=\left(1+\frac{\beta}{1-\beta}q^{2\ell}\right)^{\ell^{100}}\leqslant 1+o(1). (26)

By Equation (C.1), we have

D(x1,,x2)D(x1,,x2)\displaystyle\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})} =(1β)kss(1β+β3q2)s(1β+βq2)s\displaystyle=(1-\beta)^{k-s-s^{\prime}}\left(1-\beta+\frac{\beta}{3}q^{2\ell}\right)^{s}(1-\beta+\beta q^{2\ell})^{s^{\prime}}
(1+o(1))(1β)ks(1β+β3q2(2))s\displaystyle\leqslant(1+o(1))\cdot(1-\beta)^{k-s}\left(1-\beta+\frac{\beta}{3}q^{2(2\ell)}\right)^{s}
=(1+o(1))(1β)ks(1+β(q231))s\displaystyle=(1+o(1))\cdot(1-\beta)^{k-s}\left(1+\beta\left(\frac{q^{2\ell}}{3}-1\right)\right)^{s}
(1+o(1))exp(β(ks)+(q231)βs)\displaystyle\leqslant(1+o(1))\cdot\exp\left(-\beta(k-s)+\left(\frac{q^{2\ell}}{3}-1\right)\beta s\right)
=(1+o(1))exp(βk+q23βs).\displaystyle=(1+o(1))\cdot\exp\left(-\beta k+\frac{q^{2\ell}}{3}\beta s\right).

where in the first transition we use Equation (26), and in the fourth transition we use the fact that 1+zez1+z\leqslant e^{z}. Writing s=pk+vs=pk+v and using Equation (25) again and using vO(q(c/2+o(1)))v\leqslant O\left(q^{(c/2+o(1))\ell}\right) we have,

D(x1,,x2)D(x1,,x2)(1+o(1))exp(23βkq2+q23βv)(1+o(1))exp(O(qc/3o(1)))10.9.\frac{D^{\prime}(x_{1},\ldots,x_{2\ell})}{D(x_{1},\ldots,x_{2\ell})}\leqslant(1+o(1))\exp\left(-\frac{2}{3}\beta kq^{-2\ell}+\frac{q^{2\ell}}{3}\beta v\right)\leqslant(1+o(1))\exp\left(O(q^{-c\ell/3-o(1)})\right)\leqslant\frac{1}{0.9}.

Appendix D List Decoding Bound

In this section we prove Lemma 5.20, restated below. See 5.20

Lemma 5.20 follows directly from a generic list decoding bound of [GRS00, Theorem 15], which we state below for convenience.

Theorem D.1.

[[GRS00, Theorem 15]] Let 𝒞[Q]N\mathcal{C}\subseteq[Q]^{N} be a code with alphabet size QQ, blocklength NN, and relative distance 1γ1-\gamma. Let δ>0\delta>0 and RΣNR\in\Sigma^{N}, where Q=|Σ|Q=|\Sigma|. Suppose that C1,,CmΣNC_{1},\ldots,C_{m}\in\Sigma^{N} are distinct codewords from 𝒞\mathcal{C} that satisfy δ(R,Ci)1δ\delta(R,C_{i})\leqslant 1-\delta for all 1im1\leqslant i\leqslant m. If,

δ>1Q+(γ1Q)(11q),\delta>\frac{1}{Q}+\sqrt{\left(\gamma-\frac{1}{Q}\right)\left(1-\frac{1}{q}\right)},

then

m1(δ1/Q)2(11/Q)(γ1/Q).m\leqslant\frac{1}{(\delta-1/Q)^{2}-(1-1/Q)(\gamma-1/Q)}.

Proving Lemma 5.20 is simply a matter of translating to the notation of Theorem D.1.

Proof of Lemma 5.20.

Let Σ=𝔽q2r1\Sigma=\mathbb{F}_{q}^{2\ell-r_{1}}, let Q=span(z1,,zr1)Q=\operatorname{span}(z_{1},\ldots,z_{r_{1}}), and define a code 𝒞\mathcal{C} with alphabet Σ\Sigma by:

𝒞={(vx1,,vx2r1)(x1,,x2r1)W|vW}.\mathcal{C}=\{(v\cdot x_{1},\ldots,v\cdot x_{2\ell-r_{1}})_{(x_{1},\ldots,x_{2\ell-r_{1}})\in W}\;|\;v\in W\}.

Note that for distinct v,wWv,w\in W we have that vx=wxv\cdot x=w\cdot x for at most 1/q1/q-fraction of xWx\in W. Thus, the relative distance of CC is 1q2+r11-q^{-2\ell+r_{1}}. We would like the table TT corresponds to a word, say, RR, and f1,,fmf_{1},\ldots,f_{m} correspond to mm codewords in 𝒞\mathcal{C}, say C1,,CmC_{1},\ldots,C_{m}. A slight issue is that TT is only defined over 22\ell-dimensional subspaces of L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W], while RR has an entry for every 2r12\ell-r_{1}-tuple of points in VV. To resolve this, note that nearly every 2r12\ell-r_{1}-tuple of points combined with z1,,zr1z_{1},\ldots,z_{r_{1}} span an L𝖹𝗈𝗈𝗆[Q,W]L\in{\sf Zoom}[Q,W]. Thus, define RR as follows. If (z1,,zr1,x1,,x2r1)(z_{1},\ldots,z_{r_{1}},x_{1},\ldots,x_{2\ell-r_{1}}) are linearly independent, then let LL be the span of (z1,,zr1,x1,,x2r1)(z_{1},\ldots,z_{r_{1}},x_{1},\ldots,x_{2\ell-r_{1}}) and define

R(x1,,x2r1)=(T[L](x1),,T[L](x2r1)).R_{(x_{1},\ldots,x_{2\ell-r_{1}})}=\left(T[L](x_{1}),\ldots,T[L](x_{2\ell-r_{1}})\right).

Otherwise, define R(x1,,x2r1)R_{(x_{1},\ldots,x_{2\ell-r_{1}})} arbitrarily. Note that the fraction of tuples (x1,,x2r1)(x_{1},\ldots,x_{2\ell-r_{1}}) such that (z1,,zr1,x1,,x2r1)(z_{1},\ldots,z_{r_{1}},x_{1},\ldots,x_{2\ell-r_{1}}) are not linearly independent is at most,

i=r1+12qi1qnq2n,\sum_{i=r_{1}+1}^{2\ell}\frac{q^{i-1}}{q^{n}}\leqslant q^{2\ell-n},

so nearly all of the entries in RR correspond to table entries in TT. For the functions f1,,fmf_{1},\ldots,f_{m} we define CiC_{i} corresponding to fif_{i} by

Ci(x1,,x2r1)=(fi(x1),,fi(x2r1)).C_{i_{(x_{1},\ldots,x_{2\ell-r_{1}})}}=(f_{i}(x_{1}),\ldots,f_{i}(x_{2\ell-r_{1}})).

As each fif_{i} agrees with TT on at least β\beta-fraction of the entries, we have that RR and CiC_{i} agree on at least β\beta-fraction of the entries (x1,,x2r1)(x_{1},\ldots,x_{2\ell-r_{1}}) such that (z1,,zr1,x1,,x2r1)(z_{1},\ldots,z_{r_{1}},x_{1},\ldots,x_{2\ell-r_{1}}) are linearly independent, so

δ(R,Ci)1β(1q2n)1β2\delta(R,C_{i})\leqslant 1-\beta\cdot(1-q^{2\ell-n})\leqslant 1-\frac{\beta}{2}

for each 1im1\leqslant i\leqslant m. Finally, note that the alphabet size of 𝒞\mathcal{C} is |𝔽q2r1|=q2r1\left|\mathbb{F}_{q}^{2\ell-r_{1}}\right|=q^{2\ell-r_{1}}. To bound mm, we can apply Theorem D.1 with δ=β2q2+c2\delta=\frac{\beta}{2}\geqslant q^{-2\ell}+\frac{c}{2}, Q=q2r1Q=q^{2\ell-r_{1}}, and γ=q2+r1\gamma=q^{-2\ell+r_{1}}. We first note that the condition of Theorem D.1 is indeed satisfied,

δq2+r1+c2>q2+r1+0.\delta\geqslant q^{-2\ell+r_{1}}+\frac{c}{2}>q^{-2\ell+r_{1}}+0.

Thus Theorem D.1 implies that m4c2m\leqslant\frac{4}{c^{2}}. ∎

Appendix E Missing Proofs from Section 8

This section contains the missing proofs from Section 8, and we begin by recalling some notation. We denote by μ(𝒜)\mu(\mathcal{A}) the measure of a collcetion of subspaces 𝒜𝖦𝗋𝖺𝗌𝗌q(n,i)\mathcal{A}\subseteq{\sf Grass}_{q}(n,i), where nn and ii will always be clear from context. Furthermore, we use μX(𝒜)\mu_{X}(\mathcal{A}) to denote the measure of 𝒜\mathcal{A} restricted to the subspaces containing XX for some subspace low-dimensional subspace XX, i.e.

μX(𝒜)=|{L𝒜|XL}||{L𝖦𝗋𝖺𝗌𝗌q(n,i)|XL}|.\mu_{X}(\mathcal{A})=\frac{\left|\{L\in\mathcal{A}\;|\;X\subseteq L\}\right|}{\left|\{L\in{\sf Grass}_{q}(n,i)\;|\;X\subseteq L\}\right|}.

Likewise, when WW is a constant co-dimension subspace, we define

μW(𝒜)=|{L𝒜|LW}||{L𝖦𝗋𝖺𝗌𝗌q(n,i)|LW}|.\mu_{W}(\mathcal{A})=\frac{\left|\{L\in\mathcal{A}\;|\;L\subseteq W\}\right|}{\left|\{L\in{\sf Grass}_{q}(n,i)\;|\;L\subseteq W\}\right|}.

It will always be clear from the size of XX or WW in context which of the above definitions we are referring to. We also use

μ[X,W](𝒜)=|{L𝒜|XLW}||{L𝖦𝗋𝖺𝗌𝗌q(n,i)|XLW}|,\mu_{[X,W]}(\mathcal{A})=\frac{\left|\{L\in\mathcal{A}\;|\;X\subseteq L\subseteq W\}\right|}{\left|\{L\in{\sf Grass}_{q}(n,i)\;|\;X\subseteq L\subseteq W\}\right|},

to denote the measure of 𝒜\mathcal{A} restricted to the zoom-in of XX and the zoom-out of WW. Finally, throughout this section, for some subspace LL, and a set of constant codimension subspaces 𝒲\mathcal{W}, we will let N𝒲(L)=|{W𝒲|LW}|N_{\mathcal{W}}(L)=|\{W\in\mathcal{W}\;|\;L\subseteq W\}|.

E.1 Proof of Lemma 8.1

Recall that 𝒲\mathcal{W} is a set of m1m_{1} subspaces of codimension ss inside of VV that is tt-generic with respect to VV. For each 2(1ξ2)2\left(1-\frac{\xi}{2}\right)\ell-dimensional subspace XX and linear assignment, σ\sigma, to XX, define

pX,σ=PrWi𝒲X,σXLWi[LX,σfi|LT[L]],qX,σ=PrWi𝒲X,σXLWi[LX,σ],p_{X,\sigma}=\Pr_{\begin{subarray}{c}W_{i}\in\mathcal{W}_{X,\sigma}\\ X\subseteq L\subseteq W_{i}\end{subarray}}[L\in\mathcal{L}_{X,\sigma}\land f_{i}|_{L}\neq T[L]],\qquad q_{X,\sigma}=\Pr_{\begin{subarray}{c}W_{i}\in\mathcal{W}_{X,\sigma}\\ X\subseteq L\subset W_{i}\end{subarray}}[L\in\mathcal{L}_{X,\sigma}],

where in both probabilities XX and σ\sigma are fixed, and Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} is chosen uniformly and L𝖹𝗈𝗈𝗆[X,Wi]L\in{\sf Zoom}[X,W_{i}] is chosen uniformly. The intention behind these values is that for a fixed (X,σ)(X,\sigma), the quantity pX,σp_{X,\sigma} should reflect how much disagreement there is between the table TT and the functions fif_{i} for Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}, on subspaces LX,σL\in\mathcal{L}_{X,\sigma}, while qX,σq_{X,\sigma} should reflect the size of X,σ\mathcal{L}_{X,\sigma}. Note that if LX,σL\in\mathcal{L}_{X,\sigma} and Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}, then by definition we already have fi|XT[L]|Xσf_{i}|_{X}\equiv T[L]|_{X}\equiv\sigma. Therefore we would expect that in fact fif_{i} and T[L]T[L] also agree on LL - which is only larger than XX by ξ\xi\ell dimensions - and for most X,σX,\sigma, the value pX,σp_{X,\sigma} is small. On the other hand, for each WiW_{i}, there are at least a CC-fraction of L𝖦𝗋𝖺𝗌𝗌q(Wi,2)L\in{\sf Grass}_{q}(W_{i},2\ell) for which fi|LT[L]f_{i}|_{L}\equiv T[L], so we should also expect qX,σq_{X,\sigma} to be Ω(C)\Omega(C) for a non-trivial fraction of (X,σ)(X,\sigma). In the following claim, we formalize this intuition and show that there indeed exists an (X,σ)(X,\sigma) for which pX,σp_{X,\sigma} is small, qX,σq_{X,\sigma} is large, and additionally the set 𝒲X,σ\mathcal{W}_{X,\sigma} is large.

This idea of looking for such (X,σ)(X,\sigma) was first introduced in [IKW12] where they call these (X,σ)(X,\sigma)-excellent and was then used again in [BDN17, MZ23] to analyze lower dimensional subspace versus subspace tests, which is similar in spirit to what we are ultimately trying to show in Lemma 5.19.

Claim E.1.

There exists (X,σ)(X,\sigma) and τC2\tau\geqslant\frac{C}{2} such that:

  • m2=|𝒲X,σ|m1q10rm_{2}=|\mathcal{W}_{X,\sigma}|\geqslant\frac{m_{1}}{q^{10r\ell}}.

  • qX,στq_{X,\sigma}\geqslant\tau.

  • pX,σ<γτp_{X,\sigma}<\gamma\cdot\tau.

Proof.

Consider the following process which outputs Wi,L,X,σW_{i},L,X,\sigma such that WiW_{i} is uniform in 𝒲\mathcal{W}, L𝖦𝗋𝖺𝗌𝗌q(Wi,2)L\in{\sf Grass}_{q}(W_{i},2\ell) is uniform, X𝖦𝗋𝖺𝗌𝗌q(L,2(1ξ2))X\in{\sf Grass}_{q}\left(L,2\left(1-\frac{\xi}{2}\right)\ell\right) is uniform, and σ\sigma is the assignment of fif_{i} to XX, i.e σ:fi|X\sigma:\equiv f_{i}|_{X}.

  1. 1.

    Choose (X,σ)(X,\sigma) with probability proportional to |𝒲X,σ||\mathcal{W}_{X,\sigma}|.

  2. 2.

    Choose Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} uniformly.

  3. 3.

    Choose a 22\ell-dimensional subspace LL uniformly conditioned on XLWX\subseteq L\subseteq W.

Notice that the distribution of (Wi,L)(W_{i},L) above is equivalent to that of choosing Wi𝒲W_{i}\in\mathcal{W} uniformly and LWiL\subset W_{i} uniformly. Moreover, fi|LT[L]f_{i}|_{L}\equiv T[L] only if LX,σL\in\mathcal{L}_{X,\sigma}, as fif_{i} and T[L]T[L] must agree on XLX\subseteq L in order to agree on LL. Therefore,

𝔼X,σ[qX,σ]PrWi𝒲,LWi[fi|LT[L]]C1q2(1ξ).\mathop{\mathbb{E}}_{X,\sigma}[q_{X,\sigma}]\geqslant\Pr_{W_{i}\in\mathcal{W},L\subseteq W_{i}}[f_{i}|_{L}\equiv T[L]]\geqslant C\geqslant\frac{1}{q^{2(1-\xi)\ell}}. (27)

On the other hand,

𝔼X,σ[pX,σ]PrXL[fi|XT[L]|Xfi|LT[L]]1q2(1ξ/2).\mathop{\mathbb{E}}_{X,\sigma}[p_{X,\sigma}]\leqslant\Pr_{X\subseteq L}[f_{i}|_{X}\equiv T[L]|_{X}\;\mid\;f_{i}|_{L}\neq T[L]]\leqslant\frac{1}{q^{2(1-\xi/2)\ell}}. (28)

Here the distribution of (X,σ)(X,\sigma) is proportional to the sizes |𝒲X,σ||\mathcal{W}_{X,\sigma}| and the second inequality is by the Schwartz-Zippel lemma. Indeed, by the Schwartz-Zippel lemma, fi|Lf_{i}|_{L} and T[L]T[L] can agree on at most 1/q1/q-fraction of points zz in LL. Therefore, the middle term is bounded by the probability that 2(1ξ/2)2(1-\xi/2)\ell uniformly random, linearly independent points are all chosen in this 1/q1/q-fraction.

From a dyadic-partitioning of Equation (27), it follows that there exists a τC/2\tau\geqslant C/2 such that

PrX,σ[qX,σ[τ,2τ)]C4τlog(1/C).\Pr_{X,\sigma}\left[q_{X,\sigma}\in[\tau,2\tau)\right]\geqslant\frac{C}{4\tau\log(1/C)}. (29)

By Markov’s inequality on Equation (28)

PrX,σ[pX,σγτ]1γτq2(1ξ/2)C8τlog(1/C),\Pr_{X,\sigma}[p_{X,\sigma}\geqslant\gamma\tau]\leqslant\frac{1}{\gamma\tau q^{2(1-\xi/2)\ell}}\leqslant\frac{C}{8\tau\log(1/C)}, (30)

It follows that for at least C8tlog(1/C)\frac{C}{8t\log(1/C)}-fraction of (X,σ)(X,\sigma)’s (under the measure induced by step 1 of the sampling procedure above), we have both qX,στq_{X,\sigma}\geqslant\tau and pX,σγτp_{X,\sigma}\leqslant\gamma\tau.

Next we wish to argue that for most of these (X,σ)(X,\sigma)’s, |𝒲X,σ||\mathcal{W}_{X,\sigma}| is large. First note that the total number of (X,σ)(X,\sigma)’s is at most [dim(V)2(1ξ/2)]qq2\begin{bmatrix}{\dim(V^{\prime})}\\ {2(1-\xi/2)\ell}\end{bmatrix}_{q}q^{2\ell}. For a fixed (X,σ)(X,\sigma), the probability that it is chosen is precisely,

|𝒲X,σ|m1[dim(V)r2(1ξ/2)]q.\frac{|\mathcal{W}_{X,\sigma}|}{m}\cdot\frac{1}{\begin{bmatrix}{\dim(V)-r}\\ {2(1-\xi/2)\ell}\end{bmatrix}_{q}}.

Thus, by a union bound,

PrX,σ[|𝒲X,σ|mq10r]1q10r[n2(1ξ/2)]qq2[nr2(1ξ/2)]q1q10rq4rq21q5r.\Pr_{X,\sigma}\left[|\mathcal{W}_{X,\sigma}|\leqslant\frac{m}{q^{10r\ell}}\right]\leqslant\frac{1}{q^{10r\ell}}\cdot\frac{\begin{bmatrix}{n}\\ {2(1-\xi/2)\ell}\end{bmatrix}_{q}q^{2\ell}}{\begin{bmatrix}{n-r}\\ {2(1-\xi/2)\ell}\end{bmatrix}_{q}}\leqslant\frac{1}{q^{10r\ell}}\cdot q^{4r\ell}\cdot q^{2\ell}\leqslant\frac{1}{q^{5r\ell}}. (31)

Putting Equations (29),  (30), and (31), together, it follows that with probability at least

C4τlog(1/C)1γτq2(1ξ/2)1q5r>0,\frac{C}{4\tau\log(1/C)}-\frac{1}{\gamma\tau q^{2(1-\xi/2)\ell}}-\frac{1}{q^{5r\ell}}>0,

over (X,σ)(X,\sigma), we have, qX,στq_{X,\sigma}\geqslant\tau, pX,σγτp_{X,\sigma}\leqslant\gamma\tau, and |𝒲X,σ|m1q10r|\mathcal{W}_{X,\sigma}|\geqslant\frac{m_{1}}{q^{10r\ell}}, which establishes the claim. ∎

Taking the (X,σ)(X,\sigma) given by Claim E.1, it almost looks like Lemma 8.1 is satisfied. However, notice that while the probability of interest for the third item there looks similar to pX,σp_{X,\sigma}, it has a different distribution over LL and WiW_{i}. Indeed, there, the distribution first chooses LX,σL\in\mathcal{L}_{X,\sigma}, then Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}, whereas for pX,σp_{X,\sigma}, we are first choosing Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}, and not conditioning LWiL\subseteq W_{i} being in the set X,σ\mathcal{L}_{X,\sigma}. Intuitively, we expect something like the following to hold,

PrLX,σ,Wi𝒲X,σ[fi|LT[L]|WiL]=PrWi𝒲X,σ,LWi[fi|LT[L]|LX,σ]=pX,σqX,σγττ,\Pr_{L\in\mathcal{L}_{X,\sigma},W_{i}\in\mathcal{W}_{X,\sigma}}[f_{i}|_{L}\neq T[L]\;|\;W_{i}\supseteq L]=\Pr_{W_{i}\in\mathcal{W}_{X,\sigma},L\subseteq W_{i}}[f_{i}|_{L}\neq T[L]\;|\;L\in\mathcal{L}_{X,\sigma}]=\frac{p_{X,\sigma}}{q_{X,\sigma}}\leqslant\frac{\gamma\cdot\tau}{\tau},

and be done. These equalities are not actually true however, so the bulk of the transition from Claim E.1 to Lemma 8.1 is in formalizing this chain of equalities and converting from the distribution of pX,σp_{X,\sigma} to that required by the third item of Lemma 8.1 without losing too much.

Proof of Lemma 8.1.

Fix an (X,σ)(X,\sigma) such that Claim E.1 holds and define 𝒲X,σ\mathcal{W}_{X,\sigma} and X,σ\mathcal{L}_{X,\sigma} accordingly. Let m2=|𝒲X,σ|m1q10rm_{2}=|\mathcal{W}_{X,\sigma}|\geqslant\frac{m_{1}}{q^{10r\ell}}. In order to lower bound the measure μX(X,σ)\mu_{X}(\mathcal{L}_{X,\sigma}), we will apply Lemma 5.14 on the collection of subspaces 𝒲X,σ\mathcal{W}_{X,\sigma} with parameters j=2j=2\ell, a=2(1ξ2)a=2\left(1-\frac{\xi}{2}\right)\ell. Then the measure ν\nu over 𝖹𝗈𝗈𝗆[X,V]{\sf Zoom}[X,V] in Lemma 5.14 is precisely that obtained by choosing WiWX,σW_{i}\in W_{X,\sigma} uniformly, and then L𝖹𝗈𝗈𝗆[X,Wi]L\in{\sf Zoom}[X,W_{i}] uniformly. Thus ν\nu is precisely the distribution used to define qX,σq_{X,\sigma}, so by Lemma 5.14

μX()ν()3qs2ξm2=qX,σ3qs2ξm2τ3C6,\mu_{X}(\mathcal{L})\geqslant\nu(\mathcal{L})-\frac{3q^{\frac{s}{2}\xi\ell}}{\sqrt{m_{2}}}=q_{X,\sigma}-\frac{3q^{\frac{s}{2}\xi\ell}}{\sqrt{m_{2}}}\geqslant\frac{\tau}{3}\geqslant\frac{C}{6},

and the first two conditions of Lemma 8.1 are satisfied.

To show the third condition, it will be helpful to have in mind the bipartite graph with parts 𝒲X,σ\mathcal{W}_{X,\sigma} and X,σ\mathcal{L}_{X,\sigma} and edges (Wi,L)(W_{i},L) if LWiL\subseteq W_{i}. For each LX,σL\in\mathcal{L}_{X,\sigma} and Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} define the following degree-like quantities:

  • dL=|{Wi𝒲X,σ|WiL}|d_{L}=|\{W_{i}\in\mathcal{W}_{X,\sigma}\;|\;W_{i}\supseteq L\}|,

  • eL=|{Wi𝒲X,σ|WiL,fi|LT1[L]}|e_{L}=|\{W_{i}\in\mathcal{W}_{X,\sigma}\;|\;W_{i}\supseteq L,f_{i}|_{L}\neq T_{1}[L]\}|,

  • di=|{LX,σ|LWi}|d_{i}=|\{L\in\mathcal{L}_{X,\sigma}\;|\;L\subseteq W_{i}\}|,

  • ei=|{LX,σ|LWi,fi|LT1[L]}|e_{i}=|\{L\in\mathcal{L}_{X,\sigma}\;|\;L\subseteq W_{i},f_{i}|_{L}\neq T_{1}[L]\}|.

Also let D=|{L|XLWi,dim(L)=2}|D=|\{L\;|\;X\subseteq L\subseteq W_{i},\dim(L)=2\ell\}|, where the Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} is arbitrary (the value is the same regardless which we pick). Then 𝔼LX[dL]=m2D|X|\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X}}[d_{L}]=\frac{m_{2}D}{|\mathcal{L}_{X}|} and the probability that we are interested in bounding can be expressed as:

PrLX,σ,Wi𝒲X,σ[fi|LT1[L]|LWi]=𝔼LX,σ[eLdL].\Pr_{L\in\mathcal{L}_{X,\sigma},W_{i}\in\mathcal{W}_{X,\sigma}}[f_{i}|_{L}\neq T_{1}[L]\;|\;L\subseteq W_{i}]=\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X,\sigma}}\left[\frac{e_{L}}{d_{L}}\right].

Since qX,σtq_{X,\sigma}\geqslant t and pX,σγtp_{X,\sigma}\leqslant\gamma t, we have

qX,σm2D=LX,σdL=Wi𝒲X,σdim2Dτ,q_{X,\sigma}\cdot m_{2}\cdot D=\sum_{L\in\mathcal{L}_{X,\sigma}}d_{L}=\sum_{W_{i}\in\mathcal{W}_{X,\sigma}}d_{i}\geqslant m_{2}\cdot D\tau, (32)

and

pX,σm2D=LX,σeL=Wi𝒲X,σeim2γτD.p_{X,\sigma}\cdot m_{2}\cdot D=\sum_{L\in\mathcal{L}_{X,\sigma}}e_{L}=\sum_{W_{i}\in\mathcal{W}_{X,\sigma}}e_{i}\leqslant m_{2}\cdot\gamma\tau D. (33)

By Lemma 5.16 and the very loose bound D|X|1qξ/21q\frac{D}{|\mathcal{L}_{X}|}\approx\frac{1}{q^{\xi\ell/2}}\geqslant\frac{1}{q^{\ell}}, we have

PrLX[dL0.9𝔼L[dL]]qm2.\Pr_{L\in\mathcal{L}_{X}}\Biggl{[}d_{L}\leqslant 0.9\mathop{\mathbb{E}}_{L}[d_{L}]\Biggr{]}\leqslant\frac{q^{\ell}}{m_{2}}. (34)

Using 𝔼LX[dL]=m2D|X|\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X}}[d_{L}]=\frac{m_{2}D}{|\mathcal{L}_{X}|}, we have

𝔼LX,σ[eLdL]\displaystyle\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X,\sigma}}\left[\frac{e_{L}}{d_{L}}\right] PrLX,σ[dL0.9m2D|X|]+𝔼LX,σ[eL0.9m2D/|X|]\displaystyle\leqslant\Pr_{L\in\mathcal{L}_{X,\sigma}}\left[d_{L}\leqslant 0.9\frac{m_{2}D}{|\mathcal{L}_{X}|}\right]+\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X,\sigma}}\left[\frac{e_{L}}{0.9m_{2}D/|\mathcal{L}_{X}|}\right]
q/m2PrLX[LX,σ]+𝔼LX,σ[eL0.9m2D/|X|]\displaystyle\leqslant\frac{q^{\ell}/m_{2}}{\Pr_{L\in\mathcal{L}_{X}}[L\in\mathcal{L}_{X,\sigma}]}+\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X,\sigma}}\left[\frac{e_{L}}{0.9m_{2}D/|\mathcal{L}_{X}|}\right]
q/m2τ/3+𝔼LX,σ[eL|X|0.9m2D]\displaystyle\leqslant\frac{q^{\ell}/m_{2}}{\tau/3}+\mathop{\mathbb{E}}_{L\in\mathcal{L}_{X,\sigma}}\left[\frac{e_{L}|\mathcal{L}_{X}|}{0.9m_{2}D}\right]
3qm2τ+m2γτD0.9m2D|X||X,σ|\displaystyle\leqslant\frac{3q^{\ell}}{m_{2}\tau}+\frac{m_{2}\cdot\gamma\tau D}{0.9m_{2}D}\cdot\frac{|\mathcal{L}_{X}|}{|\mathcal{L}_{X,\sigma}|}
3qm2τ+γτ0.93τ\displaystyle\leqslant\frac{3q^{\ell}}{m_{2}\tau}+\frac{\gamma\tau}{0.9}\cdot\frac{3}{\tau}
5γ,\displaystyle\leqslant 5\gamma,

where in the second transition we use Equation (34) and in the fourth transition we use Equation (33). ∎

E.2 Proof of Lemma 8.3

Take the X,σ,X,σ,X,\sigma,\mathcal{L}^{\prime}_{X,\sigma}, and 𝒲X,σ\mathcal{W}_{X,\sigma} guaranteed by Corollary 8.2, and recall VV^{\prime} is the ambient space and δ2=ξ100\delta_{2}=\frac{\xi}{100}. As this section is more involved, we restate Lemma 8.3 as well as its setting. Recall that for a zoom-in AA and zoom-out BB such that XABVX\subseteq A\subseteq B\subseteq V^{\prime}, we write V=AV0V^{\prime}=A\oplus V_{0} and B=AVB=A\oplus V^{\star}, where VV0V^{\star}\subseteq V_{0}. Now let 𝒲[A,B]={Wi|Wi𝒲X,σ s.t AWi=WiB}\mathcal{W}^{\star}_{[A,B]}=\{W^{\star}_{i}\;|\;\exists W_{i}\in\mathcal{W}_{X,\sigma}\text{ s.t }A\oplus W^{\star}_{i}=W_{i}\cap B\}. It is clear that each Wi𝒲[A,B]W^{\star}_{i}\in\mathcal{W}^{\star}_{[A,B]} is contained inside of some Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma}, so for each WiW^{\star}_{i}, we may define fifi|Wif^{\star}_{i}\equiv f_{i}|_{W^{\star}_{i}}.

See 8.3

As a step towards Lemma 8.3, we first show Lemma E.2, which finds the basic items required for Lemma 8.3, modulo a few minor alterations.

Lemma E.2.

We can find a zoom-in AA and a zoom-out BB such that XABVX\subseteq A\subseteq B\subseteq V^{\prime}, such that the following hold.

  • dim(A)+codim(B)dim(X)+10δ2\dim(A)+\operatorname{codim}(B)\leqslant\dim(X)+\frac{10}{\delta_{2}}.

  • Letting ={LX,σ|L𝖹𝗈𝗈𝗆[A,B]}\mathcal{L}^{\prime}=\{L\in\mathcal{L}^{\prime}_{X,\sigma}\;|\;L\in{\sf Zoom}[A,B]\}, we have η=μ[A,B]()C12\eta=\mu_{[A,B]}(\mathcal{L}^{\prime})\geqslant\frac{C}{12} in 𝖹𝗈𝗈𝗆[A,B]{\sf Zoom}[A,B] and is (1,qδ2η)(1,q^{\delta_{2}\ell}\eta)-global in 𝖹𝗈𝗈𝗆[A,B]{\sf Zoom}[A,B].101010By (1,qδ2η)(1,q^{\delta_{2}\ell}\eta)-pseudo-random in 𝖹𝗈𝗈𝗆[A,B]{\sf Zoom}[A,B] we mean that \mathcal{L}^{\prime} does not increase its fractional size to qδ2ηq^{\delta_{2}\ell}\eta when restricted to any zoom-in containing AA or any zoom-out contained in BB.

Proof.

Set A0=XA_{0}=X, 0=X,σ\mathcal{L}_{0}=\mathcal{L}_{X,\sigma}, B0=VB_{0}=V, and η0=μX(X,σ)C12\eta_{0}=\mu_{X}(\mathcal{L}_{X,\sigma})\geqslant\frac{C}{12}. Now do the following.

  1. 1.

    Set i=0i=0, and initialize A0,0,B0,η0,𝒲L,0,𝒲0A_{0},\mathcal{L}_{0},B_{0},\eta_{0},\mathcal{W}_{L,0},\mathcal{W}_{0} as above.

  2. 2.

    If i\mathcal{L}_{i} is (1,qδ2ηi)(1,q^{\delta_{2}\ell}\eta_{i})-global inside of 𝖹𝗈𝗈𝗆[Ai,Bi]{\sf Zoom}[A_{i},B_{i}], then stop.

  3. 3.

    Otherwise, there exist ABA\subseteq B such that, AiABBiA_{i}\subseteq A\subseteq B\subseteq B_{i}, dim(A)+codim(B)=dim(Ai)+codim(Bi)+1\dim(A)+\operatorname{codim}(B)=\dim(A_{i})+\operatorname{codim}(B_{i})+1, and μ[A,B](i)qδ2ηi\mu_{[A,B]}(\mathcal{L}_{i})\geqslant q^{\delta_{2}\ell}\eta_{i}.

  4. 4.

    Set Ai+1=AA_{i+1}=A, Bi+1=BB_{i+1}=B, and i+1={Li+1|Ai+1LBi+1}\mathcal{L}_{i+1}=\{L\in\mathcal{L}_{i+1}\;|\;A_{i+1}\subseteq L\subseteq B_{i+1}\}.

  5. 5.

    Set ηi+1=μ[Ai+1,Bi+1](i+1)\eta_{i+1}=\mu_{[A_{i+1},B_{i+1}]}(\mathcal{L}_{i+1}).

  6. 6.

    Increment ii by 11 and return to step 22.

Suppose this process terminates on iteration jj. We claim that taking =j\mathcal{L}^{\prime}=\mathcal{L}_{j}, A=AjA=A_{j}, B=BjB=B_{j}, 𝒲L=𝒲L,j\mathcal{W}_{L}=\mathcal{W}_{L,j} for each LjL\in\mathcal{L}_{j}, and 𝒲=𝒲j\mathcal{W}^{\prime}=\mathcal{W}_{j} satisfies the requirements of the lemma.

Next notice that by construction ηi+1qδ2ηi\eta_{i+1}\geqslant q^{\delta_{2}\ell}\eta_{i}. Therefore, we perform at most log(12/C)log(qδ2)10δ2\frac{\log(12/C)}{\log(q^{\delta_{2}\ell})}\leqslant\frac{10}{\delta_{2}} iterations before stopping, so j10δ2j\leqslant\frac{10}{\delta_{2}}. By construction j\mathcal{L}_{j} is (1,qδ2ηj)(1,q^{\delta_{2}\ell}\eta_{j})-global in 𝖹𝗈𝗈𝗆[Aj,Bj]{\sf Zoom}[A_{j},B_{j}] and has fractional size ηjμX,σC12\eta_{j}\geqslant\mu_{X,\sigma}\geqslant\frac{C}{12} in 𝖹𝗈𝗈𝗆[Aj,Bj]{\sf Zoom}[A_{j},B_{j}]. Moreover, dim(Aj)+codim(Bj)=dim(X)+jdim(X)+10δ2\dim(A_{j})+\operatorname{codim}(B_{j})=\dim(X)+j\leqslant\dim(X)+\frac{10}{\delta_{2}}, so the conditions of the lemma are satisfied. ∎

Take A,BA,B and \mathcal{L}^{\prime} given by Lemma E.2. Before moving on the the straightforward derivation of Lemma 8.3, define

𝒲X,σ,L={Wi𝒲X,σ|LWi}and𝒲=L𝒲X,σ,L.\mathcal{W}_{X,\sigma,L^{\prime}}=\{W_{i}\in\mathcal{W}_{X,\sigma}\;|\;L^{\prime}\subseteq W_{i}\}\quad\text{and}\quad\mathcal{W}_{\mathcal{L}^{\prime}}=\bigcup_{L^{\prime}\in\mathcal{L}^{\prime}}\mathcal{W}_{X,\sigma,L^{\prime}}.
Proof of Lemma 8.3.

We now construct the ,𝒲,\mathcal{L}^{\star},\mathcal{W}^{\star}, and VV^{\star} that satisfy Lemma 8.3. Let VV^{\star} be a subspace such that AV=BA\oplus V^{\star}=B, set =2dim(A)\ell^{\prime}=2\ell-\dim(A), and let

={L𝖦𝗋𝖺𝗌𝗌q(V,)|LA}.\mathcal{L}^{\star}=\{L^{\star}\in{\sf Grass}_{q}(V^{\star},\ell^{\prime})\;|\;L^{\star}\oplus A\in\mathcal{L}^{\prime}\}.

For each LL^{\star}\in\mathcal{L}^{\star}, we will use LL^{\prime} to denote the corresponding subspace such that AL=LA\oplus L=L^{\prime}\in\mathcal{L}^{\prime}, and the fact that the correspondence,

LL=AL,L^{\star}\in\mathcal{L}^{\star}\longleftrightarrow L^{\prime}=A\oplus L^{\star}\in\mathcal{L}^{\prime},

is a bijection between \mathcal{L}^{\star} and \mathcal{L}^{\prime}. Recall that we abuse notation and let TT to denote both the original table on 22\ell-dimensional subspaces, as well as the new table on 𝖦𝗋𝖺𝗌𝗌q(V,){\sf Grass}_{q}(V^{\star},\ell^{\prime}), given by T[L]=T[L]|LT[L]=T[L^{\prime}]|_{L}. It will always be clear, based on the argument in T[]T[\cdot], which we are referring to. We obtain 𝒲\mathcal{W}^{\star} in a similar way as \mathcal{L}^{\star}, however, some care will be needed to ensure that it is 44-generic. First, set W~X,σ,={WiB|Wi𝒲X,σ,}\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}=\{W_{i}\cap B\;|\;W_{i}\in\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}\}, then take 𝒲~\tilde{\mathcal{W}}^{\star} to be the maximal subset of W~X,σ,\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}} that is 44-generic with codimension ss with respect to BB. Finally, set

𝒲={WiV|AWi𝒲~},\mathcal{W}^{\star}=\{W^{\star}_{i}\subseteq V^{\star}\;|\;A\oplus W^{\star}_{i}\in\tilde{\mathcal{W}}^{\star}\},

and set m3=|𝒲|m_{3}=\left|\mathcal{W}^{\star}\right|. Summarizing, we have the following chain of relations,

𝒲X,σ\displaystyle\mathcal{W}_{X,\sigma} \stackon[4pt]WiL𝒲X,σ,\stackon[4pt]B\displaystyle\mathrel{{\stackon[4pt]{$\longrightarrow$}{$\scriptscriptstyle W_{i}\supseteq L^{\prime}\in\mathcal{L}^{\prime}$}}}\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}\mathrel{{\stackon[4pt]{$\longleftrightarrow$}{$\scriptscriptstyle\cap B$}}} W~X,σ,\displaystyle\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}} \stackon[4pt]Make 4-generic\displaystyle\mathrel{{\stackon[4pt]{$\longrightarrow$}{$\scriptscriptstyle\text{Make $4$-generic}$}}} 𝒲~\displaystyle\hskip 5.69046pt\tilde{\mathcal{W}}^{\star} \stackon[4pt]Subtract subspace A𝒲\displaystyle\mathrel{{\stackon[4pt]{$\longleftrightarrow$}{$\scriptscriptstyle\text{Subtract subspace $A$}$}}}\mathcal{W}^{\star} (35)
Wi\displaystyle W_{i} \stackon[4pt]WiLWi\stackon[4pt]B\displaystyle\mathrel{{\stackon[4pt]{$\longrightarrow$}{$\scriptscriptstyle W_{i}\supseteq L^{\prime}\in\mathcal{L}^{\prime}$}}}\hskip 8.5359ptW_{i}\hskip 17.07182pt\mathrel{{\stackon[4pt]{$\longrightarrow$}{$\scriptscriptstyle\cap B$}}} WiB\displaystyle W_{i}\cap B \stackon[4pt]\displaystyle\hskip 8.5359pt\mathrel{{\stackon[4pt]{$\longrightarrow$}{$\scriptscriptstyle$}}}\hskip 5.69046pt WiB\displaystyle W_{i}\cap B \stackon[4pt]Wi s.t WiA=WiB\displaystyle\hskip 17.07182pt\mathrel{{\stackon[4pt]{$\longrightarrow$}{$\scriptscriptstyle$}}}\hskip 17.07182ptW^{\star}_{i}\text{ s.t }W^{\star}_{i}\oplus A=W_{i}\cap B (36)

which will be helpful to refer back to. The double arrow transitions are bijections, while in the single arrow transitions subspaces are being removed. The second line shows what a generic member of each set looks like, where WiW_{i} are the original subspaces in 𝒲X,σ\mathcal{W}_{X,\sigma}. We remark that we allow W~X,σ,\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}} to be a multiset. For each LL^{\star}\in\mathcal{L}^{\star}, we define

𝒲L={Wi𝒲|LWi}.\mathcal{W}^{\star}_{L^{\star}}=\{W^{\star}_{i}\in\mathcal{W}^{\star}\;|\;L^{\star}\subseteq W^{\star}_{i}\}.

It is clear from Equation (35) that 𝒲𝒲[A,B]\mathcal{W}^{\star}\subseteq\mathcal{W}_{[A,B]}. We now verify that the six properties of Lemma 8.3 hold.

Property 1. The subspaces of \mathcal{L}^{\star} are of dimension \ell^{\prime} inside VV^{\star}, and

2dim(A)2dim(X)10δ2ξ3.\ell^{\prime}\geqslant 2\ell-\dim(A)\geqslant 2\ell-\dim(X)-\frac{10}{\delta_{2}}\geqslant\frac{\xi}{3}\ell.

Also, μ()=η\mu(\mathcal{L}^{\star})=\eta is the same as the measure of \mathcal{L}^{\prime} inside 𝖹𝗈𝗈𝗆[A,B]{\sf Zoom}[A,B] due to the bijection between \mathcal{L}^{\star} and \mathcal{L}^{\prime}. Therefore which is at least ηC12\eta\geqslant\frac{C}{12} by the second part of Lemma E.2.

Property 2. Since \mathcal{L}^{\prime} does not increase its measure to qδ2ηq^{\delta_{2}\ell}\eta on any zoom-in containing AA or zoom-out inside BB, it follows that \mathcal{L}^{\star} is (1,qδ2η)(1,q^{\delta_{2}\ell}\eta)-pseudo-random.

Property 3. By construction, W~\tilde{W}^{\star} is 44-generic inside of BB. Since B=AVB=A\oplus V^{\star}, and all WiW~W_{i}\in\tilde{W}^{\star} contain AA, it follows that 𝒲\mathcal{W}^{\star} is 44-generic inside of VV^{\star}.

We verify property 4 using properties 5 and 6, so we save it for last.

Property 6. Fix an LL^{\star}\in\mathcal{L}^{\star} with corresponding LX,σL^{\prime}\in\mathcal{L}^{\prime}\subseteq\mathcal{L}_{X,\sigma}. We will first show that for all LL^{\star}\in\mathcal{L}^{\prime}, the value N𝒲(L)N_{\mathcal{W}^{\star}}(L^{\star}) is nearly the same, and in particular is nearly equal to N𝒲X,σ(L)N_{\mathcal{W}_{X,\sigma}}(L^{\prime}). This has the secondary consequence that m3N𝒲(L)m_{3}\geqslant N_{\mathcal{W}^{\star}}(L^{\star}) is large. By Lemma 5.16 applied to the 44-generic set of subspaces 𝒲\mathcal{W}^{\star} and we can conclude that Property 5 holds for most (and in particular at least one) subspaces of \mathcal{L}^{\star}. We can then conclude that the same holds for all LL^{\star}\in\mathcal{L}^{\star}.

Towards showing that all N𝒲(L)N_{\mathcal{W}^{\star}}(L^{\star}) are nearly the same, note that

N𝒲(L)=N𝒲~(L)=N𝒲~(L)|𝒲~𝒲~|,N_{\mathcal{W}^{\star}}(L^{\star})=N_{\tilde{\mathcal{W}}^{\star}}(L^{\prime})=N_{\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}}(L^{\prime})-\left|\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\right|,

and additionally

N𝒲~(L)=N𝒲X,σ,(L)=N𝒲X,σ(L),N_{\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}}(L^{\prime})=N_{\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}}(L^{\prime})=N_{\mathcal{W}_{X,\sigma}}(L^{\prime}), (37)

so

N𝒲(L)=N𝒲X,σ(L)|𝒲~𝒲~|.N_{\mathcal{W}^{\star}}(L^{\star})=N_{\mathcal{W}_{X,\sigma}}(L^{\prime})-\left|\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\right|.

Since we already have bounds on N𝒲X,σ(L)N_{\mathcal{W}_{X,\sigma}}(L^{\prime}) from the fourth part of Corollary 8.2, it is sufficient to upper bound |𝒲~𝒲~|\left|\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\right|. Since 𝒲X,σ,\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}} is 22+10δ22^{2+\frac{10}{\delta_{2}}}-generic, by Lemma 5.13, the set 𝒲~\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}} can be made 44-generic by removing at most codim(B)(22+10δ2)10δ222+10δ2\operatorname{codim}(B)\cdot(2^{2+\frac{10}{\delta_{2}}})\leqslant\frac{10}{\delta_{2}}\cdot 2^{2+\frac{10}{\delta_{2}}} of the subspaces WiBW_{i}\cap B in W~X,σ,\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}, so

|N𝒲(L)N𝒲X,σ(L)||𝒲~𝒲~|10δ222+10δ2,L𝒲.\left|N_{\mathcal{W}^{\star}}(L^{\star})-N_{\mathcal{W}_{X,\sigma}}(L^{\prime})\right|\leqslant\left|\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\right|\leqslant\frac{10}{\delta_{2}}\cdot 2^{2+\frac{10}{\delta_{2}}},\quad\quad\forall L^{\star}\in\mathcal{W}^{\star}. (38)

Using the fourth part of Corollary 8.2, we get the secondary consequence that

m30.95m2qξs10δ222+10δ2.m_{3}\geqslant 0.95\cdot m_{2}\cdot q^{\xi\ell\cdot s}-\frac{10}{\delta_{2}}\cdot 2^{2+\frac{10}{\delta_{2}}}.

Using Lemma 5.16, we can show that for at least one LL^{\star},

0.9m3qsNW(L)1.1m3qs,0.9\cdot m_{3}\cdot q^{-s\cdot\ell^{\prime}}\leqslant N_{W^{\star}}(L^{\star})\leqslant 1.1\cdot m_{3}\cdot q^{-s\cdot\ell^{\prime}}, (39)

and combining this with Equation (38) shows that Property 6 holds for all LL^{\star}\in\mathcal{L}^{\star}.

Property 5. We encourage the reader to refer back to the chain of relations in Equation (35) for this part. At a high level, we will start with a probability regarding Wi𝒲W^{\star}_{i}\in\mathcal{W}^{\star} at the right end of the chain, and gradually move leftwards and relate this to a probability regarding Wi𝒲X,σW_{i}\in\mathcal{W}_{X,\sigma} - which we have a bound on from the fourth item of Corollary 8.2. To start, note that

PrWi𝒲[fi|LT[L]|LWi]\displaystyle\Pr_{W^{\star}_{i}\in\mathcal{W}^{\star}}[f^{\star}_{i}|_{L^{\star}}\neq T[L^{\star}]\;|\;L^{\star}\subseteq W^{\star}_{i}] =PrWi𝒲[fi|LT[L]|LWi]\displaystyle=\Pr_{W^{\star}_{i}\in\mathcal{W}^{\star}}[f_{i}|_{L^{\star}}\neq T[L^{\star}]\;|\;L^{\star}\subseteq W^{\star}_{i}]
PrWiBW~[fi|LT[L]|LWiB].\displaystyle\leqslant\Pr_{W_{i}\cap B\in\tilde{W}^{\star}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}\cap B].

The first transition is simply due to the fact that fi=fi|Wif^{\star}_{i}=f_{i}|_{W^{\star}_{i}}. For the second transition we use the fact that there is a one-to-one correspondence between Wi𝒲W^{\star}_{i}\in\mathcal{W}^{\star} and Wi=AWi𝒲~W_{i}=A\oplus W^{\star}_{i}\in\tilde{\mathcal{W}}^{\star}. For this pair WiW^{\star}_{i}, WiW_{i}, we have fi=fi|Wif^{\star}_{i}=f_{i}|_{W^{\star}_{i}}. Therefore the WiW_{i} from the second probability can be sampled by first choosing WiW^{\star}_{i} according to the first distribution of the first probability, and then outputting AWiA\oplus W^{\star}_{i}. The second transition then follows.

Next we have,

PrWiBW~[fi|LT[L]|LWiB]\displaystyle\Pr_{W_{i}\cap B\in\tilde{W}^{\star}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}\cap B]
=PrWiBW~X,σ,[fi|LT[L]|LWiB,WiB𝒲~]\displaystyle=\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}\cap B,W_{i}\cap B\in\tilde{\mathcal{W}}^{\star}]
=PrWiBW~X,σ,[fi|LT[L],WiB𝒲~|LWiB]PrWiBW~X,σ,[WiB𝒲~|LWiB]\displaystyle=\frac{\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}],W_{i}\cap B\in\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B]}{\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[W_{i}\cap B\in\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B]}
PrWiBW~X,σ,[fi|LT[L]|LWiB]+2PrWiBW~X,σ,[WiBW~X,σ,𝒲~|LWiB],\displaystyle\leqslant\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}\cap B]+2\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B],

where both transitions rely on the fact that W~W~X,σ,\tilde{W}^{\star}\subseteq\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}. In the last transition, we used the fact that 1/(1δ)1+2δ1/(1-\delta)\leqslant 1+2\delta if δ1/2\delta\leqslant 1/2, thus if PrWiBW~X,σ,[WiBW~X,σ,𝒲~|LWiB]1/2\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B]\leqslant 1/2 then the inequality holds. Else, 2PrWiBW~X,σ,[WiBW~X,σ,𝒲~|LWiB]12\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B]\geqslant 1 and the expression on the third line is at most 11 so the inequality on the last transition holds trivially. We will now analyze the last two terms separately. The second term can be bounded as follows,

PrWiBW~X,σ,[WiBW~X,σ,𝒲~|LWiB]\displaystyle\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B] =|{WiB𝒲~𝒲~|LWiB}||{WiB𝒲~|LWiB}|\displaystyle=\frac{\left|\{W_{i}\cap B\in\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\;|\;L^{\prime}\subseteq W_{i}\cap B\}\right|}{|\{W_{i}\cap B\in\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\;|\;L^{\prime}\subseteq W_{i}\cap B\}|}
|𝒲~𝒲~|N𝒲X,σ(L)\displaystyle\leqslant\frac{\left|\tilde{\mathcal{W}}_{\mathcal{L}^{\prime}}\setminus\tilde{\mathcal{W}}^{\star}\right|}{N_{\mathcal{W}_{X,\sigma}}(L^{\prime})}
γ.\displaystyle\leqslant\gamma.

The first transition is evident, for the second transition note that the numerator does not decrease, while the denominator is the same (it follows from NW~X,σ,=N𝒲X,σ(L)N_{\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}=N_{\mathcal{W}_{X,\sigma}}(L^{\prime}) in Equation (37)), finally the third transition uses Equation (38) and the fact that N𝒲X,σ(L)N_{\mathcal{W}_{X,\sigma}}(L^{\prime}) is large from the fourth item of Corollary 8.2.

For the first term, note that,

PrWiBW~X,σ,[fi|LT[L]|LWiB]=PrWi𝒲X,σ,[fi|LT[L]|LWi],\displaystyle\Pr_{W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}\cap B]=\Pr_{W_{i}\in\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}],

where we use the fact that there is a one-to-one correspondence between WiBW~X,σ,W_{i}\cap B\in\tilde{W}_{X,\sigma,\mathcal{L}^{\prime}}, and Wi𝒲X,σ,W_{i}\in\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}. Also, recalling the definition of \mathcal{L}^{\prime}, we have LBL^{\prime}\subseteq B, so the conditioning in both probabilities is the same, and therefore the two probabilities are equivalent. Next, it is clear that {Wi𝒲X,σ,|LW}=𝒲X,σ,L\{W_{i}\in\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}\;|\;L^{\prime}\subseteq W^{\prime}\}=\mathcal{W}_{X,\sigma,L^{\prime}}, so using Corollary 8.2, we have,

PrWi𝒲X,σ,[fi|LT[L]|LWi]PrWi𝒲X,σ,L[fi|LT[L]]12γ.\Pr_{W_{i}\in\mathcal{W}_{X,\sigma,\mathcal{L}^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]\;|\;L^{\prime}\subseteq W_{i}]\leqslant\Pr_{W_{i}\in\mathcal{W}_{X,\sigma,L^{\prime}}}[f_{i}|_{L^{\prime}}\neq T[L^{\prime}]]\leqslant 12\gamma.

Putting everything together, we get that

PrWi𝒲[fi|LT[L]|LWi]12γ+2γ=14γ,\displaystyle\Pr_{W^{\star}_{i}\in\mathcal{W}^{\star}}[f^{\star}_{i}|_{L^{\star}}\neq T[L^{\star}]\;|\;L^{\star}\subseteq W^{\star}_{i}]\leqslant 12\gamma+2\gamma=14\gamma,

establishing property 6.

Property 4. Take an arbitrary LL^{\star}\in\mathcal{L}^{\star} with corresponding LL^{\prime}\in\mathcal{L}^{\prime} such that Equation (39) holds. We have

m3\displaystyle m_{3} qs1.1N𝒲(L)\displaystyle\geqslant\frac{q^{-s\cdot\ell^{\prime}}}{1.1}N_{\mathcal{W}^{\star}}(L^{\star})
qs1.2(N𝒲X,σ(L)10δ221+10δ2)\displaystyle\geqslant\frac{q^{-s\cdot\ell^{\prime}}}{1.2}\left(N_{\mathcal{W}_{X,\sigma}}(L^{\prime})-\frac{10}{\delta_{2}}\cdot 2^{1+\frac{10}{\delta_{2}}}\right)
qs1.2(0.95m2qξs10δ221+10δ2)\displaystyle\geqslant\frac{q^{-s\cdot\ell^{\prime}}}{1.2}\left(0.95\cdot m_{2}\cdot q^{-\xi\ell\cdot s}-\frac{10}{\delta_{2}}\cdot 2^{1+\frac{10}{\delta_{2}}}\right)
m2qs(ξ)2\displaystyle\geqslant\frac{m_{2}\cdot q^{s\cdot(\ell^{\prime}-\xi\ell)}}{2}
m2qs(dim(A)dim(X))2\displaystyle\geqslant\frac{m_{2}\cdot q^{-s\cdot(\dim(A)-\dim(X))}}{2}
m2q10s/δ22.\displaystyle\geqslant\frac{m_{2}\cdot q^{-10s/\delta_{2}}}{2}.

The first transition is by Equation (39), the second transition is by Equation (37), the third transition is by the fourth item of Corollary 8.2, the fifth transition uses the fact that =2dim(A)\ell^{\prime}=2\ell-\dim(A) and dim(X)=2(1ξ2)\dim(X)=2\left(1-\frac{\xi}{2}\right)\ell, and the last transition uses the fact that dim(A)dim(X)10δ2\dim(A)-\dim(X)\leqslant\frac{10}{\delta_{2}}. ∎

E.3 Proof of Lemma 8.5

Take ,𝒲,V\mathcal{L}^{\star},\mathcal{W}^{\star},V^{\star} given by Lemma 8.3, and recall Z={zV||μz()η|η10}Z=\{z\in V^{\star}\;|\;|\mu_{z}(\mathcal{L}^{\star})-\eta|\leqslant\frac{\eta}{10}\}. We have η=μ()\eta=\mu(\mathcal{L}^{\star}) and m3=|𝒲|m_{3}=|\mathcal{W}^{\star}|. Let us recall Lemma 8.5 below.

See 8.5

E.3.1 A Necessary Fourier Analytic Fact

We first show a Fourier Analytic fact that will be needed for the proof of Lemma 8.5. Let 𝒜𝖦𝗋𝖺𝗌𝗌q(n,j)\mathcal{A}\subseteq{\sf Grass}_{q}(n,j), for some n>qjn>q^{j}, and let η=μ(𝒜)\eta=\mu(\mathcal{A}). We define FL2(𝔽qn×j)F\in L_{2}\left(\mathbb{F}_{q}^{n\times j}\right) as follows:

F(x1,,xj)={1,ifspan(x1,,xj)𝒜,0,otherwise.F(x_{1},\ldots,x_{j})=\begin{cases}1,&\text{if}\ \operatorname{span}(x_{1},\ldots,x_{j})\in\mathcal{A},\\ 0,&\text{otherwise.}\end{cases}
Lemma E.3.

Fix a subspace W𝔽qnW\subseteq\mathbb{F}_{q}^{n}, then for any S=(s1,,sj)𝔽qn×jS=(s_{1},\ldots,s_{j})\in\mathbb{F}_{q}^{n\times j} we have,

𝔼x1,,xjW[χS(x1,,xj)]={1,ifSW,0,ifSW.\mathop{\mathbb{E}}_{x_{1},\ldots,x_{j}\in W}[\chi_{S}(x_{1},\ldots,x_{j})]=\begin{cases}1,&\text{if}\ S\subseteq W^{\perp},\\ 0,&\text{if}\ S\subsetneq W^{\perp}.\end{cases}
Proof.

If SWS\subseteq W^{\perp}, then for any xWx\in W, we have six=0s_{i}\cdot x=0 for all 1ij1\leqslant i\leqslant j, so the first case follows.

Now suppose SWS\subsetneq W^{\perp}, and without loss of generality say that s1Ws_{1}\notin W^{\perp}. We can write,

𝔼xW[χS(x)]=𝔼x1W[ωTr(x1s1)]𝔼x2,,xjW[ωi=2jTr(xisi)].\mathop{\mathbb{E}}_{x\subseteq W}[\chi_{S}(x)]=\mathop{\mathbb{E}}_{x_{1}\in W}\left[\omega^{\operatorname{Tr}(x_{1}\cdot s_{1})}\right]\mathop{\mathbb{E}}_{x_{2},\ldots,x_{j}\in W}\left[\omega^{\sum_{i=2}^{j}\operatorname{Tr}(x_{i}\cdot s_{i})}\right].

We will show that 𝔼x1W[ωTr(x1s1)]=0\mathop{\mathbb{E}}_{x_{1}\in W}\left[\omega^{\operatorname{Tr}(x_{1}\cdot s_{1})}\right]=0. Notice that it is sufficient to show that x1s1x_{1}\cdot s_{1} takes each value in 𝔽q\mathbb{F}_{q} with equal probability over uniformly random x1Wx_{1}\in W. First, since s1Ws_{1}\notin W, Prx1W[x1s1=0]=1q\Pr_{x_{1}\in W}[x_{1}\cdot s_{1}=0]=\frac{1}{q}. Next note for any α0\alpha\neq 0,

Prx1W[x1s1=1]=Prx1W[(αx1)s1=α]=Prx1W[x1s1=α].\Pr_{x_{1}\in W}[x_{1}\cdot s_{1}=1]=\Pr_{x_{1}\in W}[(\alpha x_{1})\cdot s_{1}=\alpha]=\Pr_{x_{1}\in W}[x_{1}\cdot s_{1}=\alpha].

Therefore, x1s1x_{1}\cdot s_{1} takes each of the q1q-1 nonzero values in 𝔽q\mathbb{F}_{q} with probability 1q\frac{1}{q} over uniform x1Wx_{1}\in W, and this concludes the proof. ∎

Lemma E.4.

If W𝔽qnW\subseteq\mathbb{F}_{q}^{n} has codimension ss and satisfies,

|μW(𝒜)η|0.01η,\left|\mu_{W}(\mathcal{A})-\eta\right|\geqslant 0.01\eta,

then there is a nonzero S𝔽qn×jS\in\mathbb{F}_{q}^{n\times j} such that SWS\subseteq W^{\perp} and

|F^(S)|η20qsj.\left|\widehat{F}(S)\right|\geqslant\frac{\eta}{20q^{sj}}.
Proof.

Note that, μW(𝒜)=Prx1,,xj[span(x1,,xj)𝒜|dim(span(x))=j]\mu_{W}(\mathcal{A})=\Pr_{x_{1},\ldots,x_{j}}[\operatorname{span}(x_{1},\ldots,x_{j})\in\mathcal{A}\;|\;\dim(\operatorname{span}(x))=j], so

|μW(𝒜)𝔼xW[F(x)]|jqjk.\left|\mu_{W}(\mathcal{A})-\mathop{\mathbb{E}}_{x\subseteq W}\left[F(x)\right]\right|\leqslant j\cdot q^{j-k}.

Using the Fourier decomposition of FF, we can write,

𝔼xW[F(x)]=F^(0)+0SWF^(S)𝔼xW[χS(x)]+SWF^(S)𝔼xW[χS(x)].\mathop{\mathbb{E}}_{x\subseteq W}\left[F(x)\right]=\widehat{F}(0)+\sum_{0\neq S\subseteq W^{\perp}}\widehat{F}(S)\mathop{\mathbb{E}}_{x\in W}[\chi_{S}(x)]+\sum_{\emptyset\neq S\subsetneq W^{\perp}}\widehat{F}(S)\mathop{\mathbb{E}}_{x\in W}[\chi_{S}(x)].

Using the previous inequality, and the fact that F^(0)=η\widehat{F}(0)=\eta, and Lemma E.3, we have

|μW(𝒜)η0SWF^(S)|jqjn.\left|\mu_{W}(\mathcal{A})-\eta-\sum_{0\neq S\subseteq W^{\perp}}\widehat{F}(S)\right|\leqslant j\cdot q^{j-n}.

By the triangle inequality we have,

|μW(𝒜)η||0SWF^(S)|+jqjk,\left|\mu_{W}(\mathcal{A})-\eta\right|\leqslant\left|\sum_{0\neq S\subseteq W^{\perp}}\widehat{F}(S)\right|+j\cdot q^{j-k},

and finally by the assumption in the lemma statement we have,

|0SWF^(S)|0.1ηjqjk.\left|\sum_{0\neq S\subseteq W^{\perp}}\widehat{F}(S)\right|\geqslant 0.1\cdot\eta-j\cdot q^{j-k}.

Since jqjk0.01ηj\cdot q^{j-k}\leqslant 0.01\cdot\eta and there are at most qsjq^{sj} tuples S=(s1,,sj)WS=(s_{1},\ldots,s_{j})\subseteq W^{\perp}, the result follows. ∎

E.3.2 The Proof of Lemma 8.5

For an arbitrary fixed point zVz\in V^{\star}, let DD denote the number of \ell^{\prime}-dimensional subspaces LVL\subseteq V^{\star} containing zz. We note that DD does not depend on which point zz is fixed. Also let,

z={L|zL},𝒲z={Wi𝒲|zWi},mz=|𝒲z|,N2,𝒲z(L)=|{(i,j)|WiWjL,Wi,Wj𝒲z}|,N2,𝒲(L)=|{(i,j)|WiWjL,Wi,Wj𝒲}|.\begin{split}&\mathcal{L}^{\star}_{z}=\{L\in\mathcal{L}^{\star}\;|\;z\in L\},\\ &\mathcal{W}^{\star}_{z}=\{W^{\star}_{i}\in\mathcal{W}^{\star}\;|\;z\in W^{\star}_{i}\},\\ &m_{z}=\left|\mathcal{W}^{\star}_{z}\right|,\\ &N_{2,\mathcal{W}^{\star}_{z}}(L)=|\{(i,j)\;|\;W^{\star}_{i}\cap W^{\star}_{j}\supseteq L,\;W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star}_{z}\}|,\\ &N_{2,\mathcal{W}^{\star}}(L)=|\{(i,j)\;|\;W^{\star}_{i}\cap W^{\star}_{j}\supseteq L,W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star}\}|.\end{split} (40)

Also for an arbitrary WiW^{\star}_{i} and WjW^{\star}_{j}, define

p1=PrL𝖦𝗋𝖺𝗌𝗌q(V,)[LWi]andp2=PrL𝖦𝗋𝖺𝗌𝗌q(V,)[LWiWj|zL].p_{1}=\Pr_{L\in{\sf Grass}_{q}(V^{\star},\ell^{\prime})}[L\subseteq W^{\star}_{i}]\quad\text{and}\quad p_{2}=\Pr_{L\in{\sf Grass}_{q}(V^{\star},\ell^{\prime})}[L\in W^{\star}_{i}\cap W^{\star}_{j}\;|\;z\in L]. (41)

A straightforward computation shows that p2/p12q2s/2p_{2}/p_{1}^{2}\geqslant q^{2s}/2, where recall s=codim(Wi)s=\operatorname{codim}(W^{\star}_{i}) in VV^{\star}. We start by removing all zZz\in Z that do not satisfy,

1.1qsm3mz0.9qsm3.1.1\cdot q^{-s}m_{3}\geqslant m_{z}\geqslant 0.9\cdot q^{-s}m_{3}. (42)

By Lemma 5.16, the number of zz removed is at most 2qsm3\frac{2q^{s}}{m_{3}} and is negligible, so for the remainder of the section we assume that all zZz\in Z satisfy the above inequalities.

We now define two distributions that we will later show are close to each other. The first is 𝒟1\mathcal{D}_{1}, generated by choosing zZz\in Z uniformly and Wi,Wj𝒲W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star} uniformly conditioned on zWiWjz\in W^{\star}_{i}\cap W^{\star}_{j}. The second is 𝒟1\mathcal{D}^{\prime}_{1}, generated by choosing zZz\in Z uniformly, LzL\in\mathcal{L}^{\star}_{z} uniformly, and then Wi,Wj𝒲W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star} uniformly conditioned on LWiWjL\subseteq W^{\star}_{i}\cap W^{\star}_{j}. We have

𝒟1(z,Wi,Wj)=1|Z|1|{i,j|Wi,Wj𝒲,zWiWj}|,𝒟1(z,Wi,Wj)=1|Z||{Lz|LWiWj}||z|𝔼Lz,LWiWj[1N2,𝒲(L)].\begin{split}\mathcal{D}_{1}(z,W^{\star}_{i},W^{\star}_{j})&=\frac{1}{|Z|}\cdot\frac{1}{|\{i,j\;|\;W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star},z\in W^{\star}_{i}\cap W^{\star}_{j}\}|},\\ \mathcal{D}^{\prime}_{1}(z,W^{\star}_{i},W^{\star}_{j})&=\frac{1}{|Z|}\cdot\frac{|\{L\in\mathcal{L}^{\star}_{z}\;|\;L\subseteq W^{\star}_{i}\cap W^{\star}_{j}\}|}{|\mathcal{L}^{\star}_{z}|}\cdot\mathop{\mathbb{E}}_{L\in\mathcal{L}^{\star}_{z},L\subseteq W^{\star}_{i}\cap W^{\star}_{j}}\left[\frac{1}{N_{2,\mathcal{W}^{\star}}(L)}\right].\end{split} (43)

By construction of ZZ in Lemma 8.4, we have |z|1.1ηD|\mathcal{L}^{\star}_{z}|\leqslant 1.1\eta\cdot D for all zZz\in Z. Also, since N2,𝒲(L)=N𝒲(L)2N_{2,\mathcal{W}^{\star}}(L)=N_{\mathcal{W}^{\star}}(L)^{2}, the fifth property of Lemma 8.3 yields 1.21m32p12N2,𝒲(L)0.81m32p121.21\cdot m_{3}^{2}p_{1}^{2}\geqslant N_{2,\mathcal{W}^{\star}}(L)\geqslant 0.81\cdot m_{3}^{2}p_{1}^{2}, for all LL\in\mathcal{L}^{\star}. Now, noting that p1q2sp_{1}\geqslant q^{-2s}, we have:

1|Z|11.21m32q2s𝒟1(z,Wi,Wj)\displaystyle\frac{1}{|Z|}\cdot\frac{1}{1.21m_{3}^{2}q^{-2s}}\leqslant\mathcal{D}_{1}(z,W^{\star}_{i},W^{\star}_{j}) =1|Z|1|{i,j|Wi,Wj𝒲,zWiWj}|\displaystyle=\frac{1}{|Z|}\cdot\frac{1}{|\{i,j\;|\;W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star},z\in W^{\star}_{i}\cap W^{\star}_{j}\}|}
1|Z|10.81m32q2s,\displaystyle\leqslant\frac{1}{|Z|}\cdot\frac{1}{0.81\cdot m_{3}^{2}q^{-2s}}, (44)

and

𝒟1(z,Wi,Wj)1|Z|μ[z,WiWj]()Dup21.1ηDu11.21m32p121|Z|μ[z,WiWj]()5ηm32q2s.\mathcal{D}^{\prime}_{1}(z,W^{\star}_{i},W^{\star}_{j})\geqslant\frac{1}{|Z|}\cdot\frac{\mu_{[z,W^{\star}_{i}\cap W^{\star}_{j}]}(\mathcal{L}^{\star})\cdot D_{u}\cdot p_{2}}{1.1\cdot\eta\cdot D_{u}}\cdot\frac{1}{1.21\cdot m_{3}^{2}p_{1}^{2}}\geqslant\frac{1}{|Z|}\cdot\frac{\mu_{[z,W^{\star}_{i}\cap W^{\star}_{j}]}(\mathcal{L}^{\star})}{5\cdot\eta\cdot m_{3}^{2}q^{-2s}}. (45)

By construction, |μz(z)η|η10|\mu_{z}(\mathcal{L}^{\star}_{z})-\eta|\leqslant\frac{\eta}{10} for every zZz\in Z. Call a triplet (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}) bad if

μ[z,WiWj]()45η.\mu_{[z,W^{\star}_{i}\cap W^{\star}_{j}]}(\mathcal{L}^{\star})\leqslant\frac{4}{5}\eta.

If the triplet (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}) is not bad, then by the above inequalities

𝒟1(z,Wi,Wj)1|Z|10.81m2q2s1|Z|24η25ηm2q2s6𝒟1(z,Wi,Wj).\mathcal{D}_{1}(z,W^{\star}_{i},W^{\star}_{j})\leqslant\frac{1}{|Z|}\cdot\frac{1}{0.81\cdot m^{2}q^{-2s}}\leqslant\frac{1}{|Z|}\cdot\frac{24\eta}{25\cdot\eta\cdot m^{2}q^{-2s}}\leqslant 6\mathcal{D}^{\prime}_{1}(z,W^{\star}_{i},W^{\star}_{j}). (46)

We start by showing that there are very few bad triplets.

Claim E.5.

For each zz, the number of i,ji,j such that

μ[z,WiWj]()45η,\mu_{[z,W^{\star}_{i}\cap W^{\star}_{j}]}(\mathcal{L}^{\star})\leqslant\frac{4}{5}\eta,

is at most 106q4sη2m3\frac{10^{6}q^{4s\ell}}{\eta^{2}}m_{3}. Additionally, for every Wi,Wj𝒲W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star}, we have

μWiWj(Z)0.9μ(Z).\mu_{W^{\star}_{i}\cap W^{\star}_{j}}(Z)\geqslant 0.9\mu(Z).
Proof.

Fix a point zz, let FzF_{z} be the restriction of FF to the zoom-in of zz, where F(x1,,x)=1F(x_{1},\ldots,x_{\ell^{\prime}})=1 if span(x1,,x)\operatorname{span}(x_{1},\ldots,x_{\ell^{\prime}})\in\mathcal{L}^{\star} and 0 otherwise. Let η=μz()\eta^{\prime}=\mu_{z}(\mathcal{L}^{\star}). For any i,ji,j satisfying the inequality of the lemma,

|μ[z,WiWj]()η|η20.\left|\mu_{[z,W^{\star}_{i}\cap W^{\star}_{j}]}(\mathcal{L}^{\star})-\eta^{\prime}\right|\geqslant\frac{\eta^{\prime}}{20}.

We can then apply Lemma E.4 to the zoom-in on zz. By Lemma E.4, if (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}) is bad then there must be S=(s1,,s1)S=(s_{1},\ldots,s_{\ell^{\prime}-1}) such that span(s1,,s1)(WiWj)\operatorname{span}(s_{1},\ldots,s_{\ell^{\prime}-1})\subseteq\left(W^{\star}_{i}\cap W^{\star}_{j}\right)^{\perp} and,

|Fz^(S)|η400q2s(1).|\widehat{F_{z}}(S)|\geqslant\frac{\eta^{\prime}}{400q^{2s(\ell^{\prime}-1)}}. (47)

Since by Parseval’s inequality the sum of |Fz^(S)|2\left|\widehat{F_{z}}(S)\right|^{2} is at most Fz221\|F_{z}\|_{2}^{2}\leqslant 1, there are at most 160000q4sη2\frac{160000q^{4s\ell^{\prime}}}{\eta^{\prime 2}} tuples SS satisfying (47). Now consider a bipartite graph where the left side consists of these tuples S=(s1,,s1)S=(s_{1},\ldots,s_{\ell^{\prime}-1}), the right side consists of WiWjW^{\star}_{i}\cap W^{\star}_{j}, and the edges are between pairs that satisfy

span(s1,,s1)(WiWj).\operatorname{span}(s_{1},\ldots,s_{\ell^{\prime}-1})\subseteq\left(W^{\star}_{i}\cap W^{\star}_{j}\right)^{\perp}.

It follows that the number of edges in this graph is an upper bound on the number of bad triples containing zz. Since 𝒲\mathcal{W}^{\star} is 44-generic, we have

(WiWj)(WiWj)={0}(W^{\star}_{i}\cap W^{\star}_{j})^{\perp}\cap(W^{\star}_{i^{\prime}}\cap W^{\star}_{j^{\prime}})^{\perp}=\{0\}

for all i,j,i,ji,j,i^{\prime},j^{\prime} distinct. Therefore, any two neighbours of a vertex on the left must either have their ii or jj be equal, and hence the maximum degree of a vertex on the left side is at most 2m32m_{3}. As a result, the graph has at most 2m3160000q4sη2106q4sη2m32m_{3}\cdot\frac{160000q^{4s\ell}}{\eta^{2}}\leqslant\frac{10^{6}q^{4s\ell}}{\eta^{\prime 2}}m_{3} edges, where we also use that η0.9η\eta^{\prime}\geqslant 0.9\eta. This completes the proof of the first assertion of the claim.

For the second part of the lemma, note that μ(Z¯)q2\mu(\overline{Z})\leqslant\frac{q^{-\ell^{\prime}}}{2} by Lemma 8.4. Therefore, for any WiWjW^{\star}_{i}\cap W^{\star}_{j}, we have,

μWiWj(Z¯)q2sq2.\mu_{W^{\star}_{i}\cap W^{\star}_{j}}(\overline{Z})\leqslant q^{2s}\cdot\frac{q^{-\ell^{\prime}}}{2}.

It follows that,

μWiWj(Z)1q2sq20.9μ(Z).\mu_{W^{\star}_{i}\cap W^{\star}_{j}}(Z)\geqslant 1-q^{2s}\cdot\frac{q^{-\ell^{\prime}}}{2}\geqslant 0.9\mu(Z).

Lemma E.6.

Let EE be any event defined with respect to (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}). Then,

𝒟1(E)6𝒟1(E)+γ.\mathcal{D}_{1}(E)\leqslant 6\mathcal{D}^{\prime}_{1}(E)+\gamma.
Proof.

If the triple (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}) is not bad, then 𝒟1(z,Wi,Wj)6𝒟1(z,Wi,Wj)\mathcal{D}_{1}(z,W^{\star}_{i},W^{\star}_{j})\leqslant 6\mathcal{D}^{\prime}_{1}(z,W^{\star}_{i},W^{\star}_{j}). Otherwise, we can use the generic bound 𝒟1(z,Wi,Wj)10.81m32q2s\mathcal{D}_{1}(z,W^{\star}_{i},W^{\star}_{j})\leqslant\frac{1}{0.81\cdot m_{3}^{2}q^{-2s}}, which can be obtained from Equation (E.3.2). By the bound on the number of bad triples per zz in Claim E.5, it follows that

𝒟1(E)\displaystyle\mathcal{D}_{1}(E) 6𝒟1(E)+|Z|106q4sη2m31|Z|0.81m32q2s\displaystyle\leqslant 6\mathcal{D}^{\prime}_{1}(E)+|Z|\cdot\frac{10^{6}q^{4s\ell}}{\eta^{2}}m_{3}\frac{1}{|Z|\cdot 0.81\cdot m_{3}^{2}q^{-2s}}
=6𝒟1(E)+107q4s+2sη2m3\displaystyle=6\mathcal{D}^{\prime}_{1}(E)+\frac{10^{7}q^{4s\ell+2s}}{\eta^{2}\cdot m_{3}}
6𝒟1(E)+γ.\displaystyle\leqslant 6\mathcal{D}^{\prime}_{1}(E)+\gamma.

Note that in the last transition we are using the fact that m3m_{3} is large by the fourth property in Lemma 8.3. ∎

Now let 𝒟2\mathcal{D}_{2} be the distribution obtained by choosing Wi,Wj𝒲W^{\star}_{i},W^{\star}_{j}\in\mathcal{W} uniformly, and then choosing zWiWjZz\in W^{\star}_{i}\cap W^{\star}_{j}\cap Z uniformly. We have

𝒟2(z,Wi,Wj)=1m321|WiWjZ|.\mathcal{D}_{2}(z,W^{\star}_{i},W^{\star}_{j})=\frac{1}{m_{3}^{2}}\cdot\frac{1}{|W^{\star}_{i}\cap W^{\star}_{j}\cap Z|}.

Using essentially the same proof, we get the following lemma.

Lemma E.7.

Let EE be any event defined with respect to (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}). Then,

𝒟2(E)2𝒟1(E).\mathcal{D}_{2}(E)\leqslant 2\mathcal{D}_{1}(E).
Proof.

We apply Claim E.4 where V=VV=V^{\star} and =Z\mathcal{L}=Z. By Claim E.5, we have μWiWj(Z)>0.9μ(Z)\mu_{W^{\star}_{i}\cap W^{\star}_{j}}(Z)>0.9\mu(Z) for all i,ji,j, or equivalently

|WiWjZ|0.9|Z|q2s.|W^{\star}_{i}\cap W^{\star}_{j}\cap Z|\geqslant 0.9\cdot|Z|q^{-2s}.

Thus, for all i,ji,j and all zz,

𝒟2(z,Wi,Wj)10.9|Z|q2s1m3221.21|Z|1m32q2s2D1(z,Wi,Wj),\mathcal{D}_{2}(z,W^{\star}_{i},W^{\star}_{j})\leqslant\frac{1}{0.9|Z|\cdot q^{-2s}}\cdot\frac{1}{m_{3}^{2}}\leqslant\frac{2}{1.21|Z|}\cdot\frac{1}{m_{3}^{2}q^{-2s}}\leqslant 2D_{1}(z,W^{\star}_{i},W^{\star}_{j}),

where we use Equation (E.3.2) for the third transition. It follows that

𝒟2(E)2𝒟1(E).\mathcal{D}_{2}(E)\leqslant 2\mathcal{D}_{1}(E).\qed

We are now ready to prove Lemma 8.5.

Proof of Lemma 8.5.

By the sixth property in Lemma 8.3, for every LL\in\mathcal{L}^{\star}, we have

PrWiL,Wi𝒲[fi|LT1[L]]14γ.\Pr_{W_{i}^{\star}\supseteq L,W^{\star}_{i}\in\mathcal{W}^{\star}}[f_{i}|_{L}\neq T_{1}[L]]\leqslant 14\gamma.

Let EE denote the event over (z,Wi,Wj)(z,W^{\star}_{i},W^{\star}_{j}) that fi(z)fj(z)f_{i}(z)\neq f_{j}(z). It follows that,

𝒟1(E)PrWi,WjLWi,Wj𝒲[fi|LT1[L]fj|LT1[L]]2PrWiL,Wi𝒲[fi|LT1[L]]28γ.\mathcal{D}_{1}(E)\leqslant\Pr_{\begin{subarray}{c}W_{i}^{\star},W^{\star}_{j}\supseteq L\\ W^{\star}_{i},W^{\star}_{j}\in\mathcal{W}^{\star}\end{subarray}}[f_{i}|_{L}\neq T_{1}[L]\;\lor\;f_{j}|_{L}\neq T_{1}[L]]\leqslant 2\cdot\Pr_{W_{i}^{\star}\supseteq L,W^{\star}_{i}\in\mathcal{W}^{\star}}[f_{i}|_{L}\neq T_{1}[L]]\leqslant 28\gamma.

Putting Lemmas E.6 and  E.7 together,

𝒟2(E)2(6𝒟1(E)+γ)=338γ,\mathcal{D}_{2}(E)\leqslant 2(6\mathcal{D}_{1}(E)+\gamma)=338\gamma,

proving the first part of Lemma 8.5.

For the second part, recall from the second part of Claim E.5 that for every pair i,ji,j, we have

|WiWjZ|0.9|Z|q2s0.81|V|q2s=0.81|WiWj|.\left|W^{\star}_{i}\cap W^{\star}_{j}\cap Z\right|\geqslant 0.9\cdot\left|Z\right|q^{-2s}\geqslant 0.81\left|V^{\star}\right|q^{-2s}=0.81\cdot\left|W^{\star}_{i}\cap W^{\star}_{j}\right|.\qed