Near Optimal Alphabet-Soundness Tradeoff PCPs
Abstract
We show that for all , for sufficiently large prime power , for all , it is NP-hard to distinguish whether a -Prover--Round projection game with alphabet size has value at least , or value at most . This establishes a nearly optimal alphabet-to-soundness tradeoff for -query PCPs with alphabet size , improving upon a result of Chan [Cha16]. Our result has the following implications:
-
1.
Near optimal hardness for Quadratic Programming: it is NP-hard to approximate the value of a given Boolean Quadratic Program within factor under quasi-polynomial time reductions. This result improves a result of Khot and Safra [KS13] and nearly matches the performance of the best known approximation algorithm [Meg01, NRT99, CW04] that achieves a factor of .
-
2.
Bounded degree -CSP’s: under randomized reductions, for sufficiently large , it is NP-hard to approximate the value of -CSPs in which each variable appears in at most constraints within factor , improving upon a recent result of Lee and Manurangsi [LM23].
- 3.
1 Introduction
The PCP theorem is a fundamental result in theoretical computer science with many equivalent formulations [FGL+91, AS03, ALM+92]. One of the formulations asserts that there exists such that given a satisfiable -SAT formula , it is NP-hard to find an assignment that satisfies at least fraction of the constraints. The PCP theorem has a myriad of applications within theoretical computer science, and of particular interest to this paper are applications of PCP to hardness of approximation.
The vast majority of hardness of approximation result are proved via reductions from the PCP theorem above. Oftentimes, to get a strong hardness of approximation result, one must first amplify the basic PCP theorem above into a result with stronger parameters [Hås01, Hås96, Fei98, KP06] (see [Tre14] for a survey). To discuss these parameters, it is often convenient to view the PCP through the problem of -Prover--Round Games, which we define next.111Strictly speaking, the notion below is referred to in the literature as projection -Prover--Round games. We omit the more general definition as we do not discuss non-projection games in this paper.
Definition 1.1.
An instance of -Prover--Round Games consists of a bipartite graph , alphabets and and a collection of constraints , which for each edge specifies a constraint map .
-
1.
The alphabet size of is defined to be .
-
2.
The value of is defined to be the maximum fraction of edges that can be satisfied by any assignment. That is,
The combinatorial view of -Prover--Round Games has its origins in an equivalent, active view in terms of a game between a verifier and two all powerful provers, which is sometimes more intuitive. The verifier and the two provers have access to an instance of -Prover--Round Games, and the provers may agree beforehand on a strategy; after this period they are not allowed to communicate. The verifier then picks a random edge, , from the -Prover--Round game, sends to the first prover, sends to the second prover, receives a label in response from each one of them, and finally checks that the labels satisfy the constraint . If so, then the verifier accepts. It is easy to see that the value of the -Prover--Round game is equal to the acceptance probability of the verifier under the best strategy of the provers. This view will be useful for us later.
In the language of -Prover--Round Games, the majority of hardness of approximation results are proved by combining the basic PCP theorem [FGL+91, AS03, ALM+92] with Raz’s parallel repetition theorem [Raz98], which together imply the following result:
Theorem 1.2.
There exists such that for sufficiently large , given a -Prover--Round game with alphabet size , it is NP-hard to distinguish between the following two cases:
-
1.
YES case: .
-
2.
NO case: .
For many applications, one only requires that the soundness error of the PCP is small. Namely, that is arbitrarily small in the “NO case”. For certain applications however, more is required: not only must the soundness error be small – but it must also be small in terms of the alphabet size. The tradeoff between the soundness error of the PCP and the alphabet size of the PCP is the main focus of this paper.
With respect to this tradeoff, it is clear that the best result one may hope for in Theorem 1.2 is since a random assignment to satisfies, in expectation, at least fraction of the constraints. In terms of results, combining the PCP theorem with Raz’s parallel repetition theorem gives that is an absolute, but tiny constant. Towards a stronger tradeoff, Khot and Safra [KS13] showed that Theorem 1.2 holds for with imperfect completeness (i.e., instead of in the YES case). The result of Khot and Safra was improved by Chan [Cha16], who showed (using a completely different set of techniques) that Theorem 1.2 holds for , again with imperfect completeness.
1.1 Main Results
In this section we explain the main results of this paper.
1.1.1 Near Optimal Alphabet vs Soundness Tradeoff
The main result of this work improves upon all prior results, and shows that one may take in Theorem 1.2, again with imperfect completeness. Formally, we show:
Theorem 1.3.
For all , for sufficiently large , given a -Prover--Round game , it is NP-hard to distinguish between the following two cases:
-
1.
YES case: .
-
2.
NO case: .
Theorem 1.3 shows a near optimal tradeoff between the alphabet of a PCP and the alphabet of a PCP, improving upon the result of Chan [Cha16]. Moreover, Theorem 1.3 has several applications to combinatorial optimization problems, which we discuss below. We remark that most of these applications require additional features from the instances produced in Theorem 1.3 which we omit from its formulation for the sake of clarity. For instance, one application requires a good tradeoff between the size of the instance and the size of the alphabet, which our construction achieves (see the discussion following Theorem 1.4). Other applications require the underlying constraint graph to be bounded-degree bi-regular graph, which our construction also achieves (after mild modifications; see Theorem 7.1).
1.1.2 Application: NP-Hardness of Approximating Quadratic Programs
Theorem 1.3 has an application to the hardness of approximating the value of Boolean Quadratic Programming, as we explain next.
An instance of Quadratic programming consists of a quadratic form where for all , and one wishes to maximize over . This problem is known to have an approximation algorithm [Meg01, NRT99, CW04], and is known to be quasi-NP-hard to approximate within factor [ABH+05, KS13]. That is, unless NP has a quasi-polynomial time algorithm, no polynomial time algorithm can approximate Quadratic Programming to within factor . As a first application of Theorem 1.3, we improve the hardness result of Khot and Safra:
Theorem 1.4.
It is quasi-NP-hard to approximate Quadratic Programming to within a factor of .
Theorem 1.4 is proved via a connection between -Prover--Round Games and Quadratic Programming due to Arora, Berger, Hazan, Kindler, and Safra [ABH+05]. This connections requires a good tradeoff between the alphabet size, the soundness error of the PCP, and the size of the PCP. Fortunately, the construction in Theorem 1.4 has a sufficiently good tradeoff between all of these parameters: letting be the size of the instance, the alphabet size can be taken to be and the soundness error can be taken to be . 222We remark that the result of Chan [Cha16] does not achieve a good enough trade-off between the alphabet size and the instance size due to the use of the long-code, and therefore it does not yield a strong inapproximability result for Quadratic Programming.
Relevance to the sliding scale conjecture:
It is worth noting that using our techniques, we do not know how to achieve soundness error that is smaller than inversely poly-logarithmic in the instance size. In particular, our techniques have no bearing on the sliding scale conjecture, which is concerned with getting soundness error that is inversely polynomial in the instance size. This seems to be a bottleneck of any PCP construction that is based on the covering property. In fact, assuming ETH, any quasi-polynomial PCP construction achieving soundness error, say, would necessarily need to have almost polynomial alphabet size (since the reduction to Quadratic Solvability would give an algorithm that runs roughly in time exponential in the alphabet size), which is the opposite of what our techniques give. With this in mind, we would like to mention a closely related, recent conjecture made in [CDGT22], which is a sort of a mixture between -to- games and the sliding scale conjecture. This conjecture is motivated by improved hardness results for densest -subgraph style problems, and focuses on the relation between the instance size and the soundness error (allowing the alphabet to be quite large). It may be possible that the ideas from the current paper can help make progress towards this conjecture.
1.1.3 Application: NP-hardness of Approximating Bounded Degree -CSPs
Theorem 1.3 has an application to the hardness of approximating the value of -CSPs with bounded degree, as we explain next.
An instance of -CSP, say , consists of a set of variables , a set of constraints and an alphabet . Each constraint in has the form where is a predicate (which may be different in distinct constraints). The degree of the instance is defined to be the maximum, over variables , of the number of constraints that appears in. The goal is to find an assignment that satisfies as many of the constraints as possible.
There is a simple approximation algorithm for the -CSP problem for instances with degree at most . Lee and Manurangsi proved a nearly matching hardness of approximation result assuming the Unique-Games Conjecture [LM23]. Unconditionally, they show the problem to be NP-hard to approximate within factor under randomized reductions.
Using the ideas of Lee and Manurangsi, our main result implies a nearly matching NP-hardness result for bounded degree -CSPs:
Theorem 1.5.
For all , for sufficiently large , approximating the value of -CSPs with degree at most within factor is NP-hard under randomized reductions.
As in [LM23], Theorem 1.5 has a further application to finding independent sets in claw free graphs. A -claw is the vertex graph with a center vertex which is connected to all other -vertices and has no other edges; a graph is said to be -claw free if does not contain an induced -claw graph. There is a polynomial time approximation algorithm for approximating the size of the largest independent set in a given -claw free graph within factor [Ber00, TW23], and a quasi-polynomial time approximation algorithm within factor [CGM13]. As in [LM23], using ideas from [DFRR23] Theorem 1.5 implies that for all , for sufficiently large , it is NP-hard (under randomized reductions) to approximate the size of the largest independent set in a given -claw free graph within factor . This improves upon the result of [LM23] who showed that the same result holds assuming the Unique-Games Conjecture.
1.1.4 Application: NP-hardness of Approximating Connectivity Problems
Using ideas of Laekhanukit [Lae14] and the improvements by Manurangsi [Man19], Theorem 1.3 implies improved hardness of approximation results for several graph connectivitiy problems. More specifically, Theorem 1.3 combined with the results of [Man19] implies improvements to each one of the results outlined in table in [Lae14] by a factor of in the exponent - with the exception of Rooted--Connectivity on directed graphs where a factor of improvement is already implied by [Man19]. We briefly discuss the Rooted -Connectivity Problem, but defer the reader to [Lae14] for a detailed discussion of the remaining graph connectivity problems.
In the Rooted -Connectivity problem there is a graph , edge costs , a root vertex and a set of terminals . The goal is to find a sub-graph of smallest cost that for each , has at least vertex disjoint paths from to . The problem admits trivial approximation algorithm (by applying minimum cost -flow algorithm for each vertex in ), as well as an approximation algorithm [Nut12].
Using the ideas of [Lae14], Theorem 1.3 implies the following improved hardness of approximation results:
Theorem 1.6.
For all , for sufficiently large it is NP-hard to approximate the Rooted--Connectivity problem on undirected graphs to within a factor of , the Vertex-Connectivity Survivable Network Design Problem with connectivity parameters at most to within a factor of , and the Vertex-Connectivity -Route Cut Problem to within a factor of .
We remark that in [CCK08], a weaker form of hardness for the Vertex-Connectivity Survivable Network problem is proved. More precisely, they show an integrality gap for the set-pair relaxation of the problem. Our hardness result of improves upon it, showing that (unless PNP) no relaxation can yield a better than factor approximation algorithm.
1.2 Our Techniques
Theorem 1.3 is proved via a composition of an Inner PCP and an Outer PCP. Both of these components incorporate ideas from the proof of the -to- Games Theorem. The outer PCP is constructed using smooth parallel repetition [KS13, KMS17] while the inner PCP is based on the Grassmann graph [KMS17, DKK+18, DKK+21, KMS23].
The novelty in this current paper, in terms of techniques, is twofold. First, we must consider a Grassmann test in a different regime of parameters (as otherwise we would not be able to get a good alphabet to soundness tradeoff) and in a regime of much lower soundness error. These differences complicate matters considerably. Second, our soundness analysis is more involved than that of the -to--Games Theorem. As is the case in [KMS17, DKK+18, DKK+21, KMS23], we too use global hyperconractivity, but we do so more extensively. We also require quantitatively stronger versions of global hypercontractivity over the Grasssmann graph which are due to [EKL24]. In addition, our analysis incorporates ideas from the plane versus plane test and direct product testing [RS97, IKW12, MZ23], from classical PCP theory [KS13], as well as from error correcting codes [GRS00]. All of these tools are necessary to prove our main technical statement – Lemma 1.7 below – which is a combinatorial statement that may be of independent interest.
We now elaborate on each one of the components separately.
1.2.1 The Inner PCP
Our Inner PCP is based on the subspace vs subspace low degree test. Below, we first give a general overview of the objective in low-degree testing. We then discuss the traditional notion of soundness as well as a non-traditional notion of soundness for low-degree tests. Finally, we explain the low-degree test used in this paper, the notion of soundness that we need from it, and the way that this notion of soundness is used.
Low degree tests in PCPs.
Low degree tests have been have a vital component in PCPs since their inception, and much attention has been devoted to improving their various parameters. The goal in low-degree testing is to encode a low-degree function via a table (or a few tables) of values, in a way that allows for local testing. Traditionally, one picks a parameter (which is thought of as a constant and is most often just ) and encodes the function by the table of restrictions of to -dimensional affine subspaces of . For the case , the test associated with this encoding is known as the Plane vs Plane test [RS97]. The Plane vs Plane test proceeds by picking two planes , intersecting on a line, and then checking that and agree on . It is easy to see that the test has perfect completeness, namely that a valid table of restrictions passes the test with probability . In the other direction, the soundness error of the test – which is a converse type statement – is much less clear (and is crucial towards applications in PCP). In the context of the Plane vs Plane test, it is know that if a table , that assigns to each plane a degree function, passes the Plane vs Plane test with probability (where is a small absolute constant), then there is a degree function such that on at least fraction of the planes.
Nailing down the value of the constant for which soundness holds is an interesting open problem which is related to soundness vs alphabet size vs instance size tradeoff in PCPs [MR10, BDN17, MZ23]. Currently, the best known analysis for the Plane vs Plane test [MR10] shows that one may take . Better analysis is known for higher dimensional encoding [BDN17, MZ23], and for the -dimensional version of it a near optimal soundness result is known [MZ23].
Low degree tests in this paper.
In the context of the current paper, we wish to encode linear functions , and we do so by the subspaces encoding. Specifically, we set integer parameters , and encode the function using the table of the restrictions of to all -dimensional linear subspaces of , and the table of the restrictions of to all -dimensional linear subspaces of . The test we consider is the natural inclusion test:
-
1.
Sample a random -dimensional subspace and a random -dimensional subspace .
-
2.
Read , and accept if they agree on .
As is often the case, the completeness of the test – namely the fact that valid tables pass the test with probability – is clear. The question of most interest then is with regards to the soundness of the test. Namely, what is the smallest such that any two tables and that assign linear functions to subspaces and pass the test with probability , must necessarily “come from” a legitimate linear function ?
Traditional notion of soundness.
As the alphabet vs soundness tradeoff is key to the discussion herein, we begin by remarking that the alphabet size of the above encoding is (since there are distinct linear functions on a linear space of dimension over ). Thus, ideally we would like to show that the soundness error of the above test is . Alas, this is false. Indeed, it turns out that one may construct assignments that pass the test with probability at least that do not have significant correlation with any linear function :
-
1.
Taking randomly by assigning to each subspace a random linear function, one can easily see that the test passes with probability .
-
2.
Taking linear subspaces of co-dimension randomly, and a random linear function for each , one may choose and as follows. For each , pick a random such that (if such exists) and assign . For each , pick a random such that (if such exists) and assign . Taking randomly, one sees that with constant probability has many possible ’s, has many possible ’s and furthermore there is at least one that is valid for both of them. With probability this common is chosen for both and , and in this case, the test on passes. It follows that, in expectation, pass the test with probability .
In light of the above, it makes sense that the best possible alphabet vs soundness tradeoff we may achieve with the subspace encoding is by taking . Such a setting of the parameters would give alphabet size and (possibly) soundness error . There are several issues with this setting however. First, this tradeoff is not good enough for our purposes (which already rules out this setting of parameters). Second, we do not know how to prove that the soundness error of the test is (the best we can do is quadratically off and is ). To address both of these issues, we must venture beyond the traditional notion of soundness.
Non-traditional notion of soundness.
The above test was first considered in the context of the -to- Games Theorem, wherein one takes and . In this setting, the test is not sound in the traditional sense; instead, the test is shown to satisfy a non-standard notion of soundness, which nevertheless is sufficient for the purposes of constructing a PCP. More specifically, in [KMS23] it is proved that for all there is such that for sufficiently large and for tables as above, there are subspaces with and a linear function such that
We refer to the set
as the zoom in of and zoom out of . While this result is good for the purposes of -to- Games, the dependency between and (and thus, between the soundness and the alphabet size) is still not good enough for us.
Our low-degree test.
It turns out that the proper setting of parameters for us is where is a small constant. With these parameters, we are able to show that for (where is a vanishing function of ), if , pass the test with probability at least , then there are subspaces with , and a linear function such that
Working in the very small soundness regime of entails with it many challenges, however. First, dealing with such small soundness requires us to use a strengthening of the global hypercontractivity result of [KMS23] in the form of an optimal level inequality due to Evra, Kindler and Lifshitz [EKL24]. Second, in the context of [KMS23], could be any function of (and indeed it ends up being a polynomial function of ). In the context of the current paper, it is crucial that , as opposed to, say, . The reason is that, as we are dealing with very small , the result would be trivial for and not useful towards the analysis of the PCP (as then would be below the threshold which represents the agreement a random linear function has with ).
1.2.2 Getting List Decoding Bounds
As is usually the case in PCP reductions, we require a list decoding version for our low-degree test. Indeed, using a standard argument we are able to show that in the setting that and , there is such that for at least fraction of subspaces of dimension , there exists a subspace with co-dimension at most and , as well as a linear function , such that
(1) |
This list decoding version theorem alone is not enough. In our PCP construction, we compose the inner PCP with an outer PCP (that we describe below), and analyzing the composition requires decoding global linear functions (from a list decoding version theorem as above) in a coordinated manner between two non communicating parties. Often times, the number of possible global functions that may be decoded is constant, in which case randomly sampling one among them often works. This is not the case for us, though: if and are distinct zoom-in and zoom-out pairs for which there are linear functions and satisfying (1), then the functions and could be completely different. Thus, to achieve a coordinated decoding procedure, we must:
-
1.
Facilitate a way for the two parties to agree on a zoom-in and zoom-out pair with noticeable probability.
-
2.
Show that for each there are at most functions for which
The second item is precisely the reason we need to be ; any worse dependency, such as would lead to the second item being false. We also remark that the number of functions being is important to us as well. There is some slack in this bound, but a weak quantitative bound such as would have been insufficient for some of our applications. Luckily, such bounds can be deduced from [GRS00] for the case of linear functions.333In the case of higher degree functions (even quadratic functions) some bounds are known [Gop13, BL15] but they would not have been good enough for us.
We now move onto the first item, in which we must facilitate a way for two non-communicating parties to agree on a zoom-in and zoom-out pair . It turns out that agreeing on the zoom-in can be delegated to the outer PCP, and we can construct a sound outer PCP game in which the two parties are provided with a coordinated zoom-in . This works because in our list decoding theorem, the fraction of zoom-ins that work is significant. Coordinating zoom-outs is more difficult, and this is where much of the novelty in our analysis lies.
1.2.3 Coordinating Zoom-outs
For the sake of simplicity and to focus on the main ideas, we ignore zoom-ins for now and assume that the list decoding statement holds with no . Thus, the list decoding theorem asserts that there exists a zoom-out of constant co-dimension on which there is a global linear function. However, there could be many such zoom-outs , say and say all of them were of co-dimension . If the number were sufficiently large – say at least fraction of all co-dimension subspaces – then we would have been able to coordinate them in the same way as we coordinate zoom-ins. If the number were sufficiently small – say , then randomly guessing a zoom-out would work well enough. The main issue is that the number is intermediate, say .
This issue had already appeared in [KMS17, DKK+18]. Therein, this issue is resolved by showing that if there are at least zoom-outs of co-dimension , and linear functions on respectively such that
for all , then there exists a zoom out of co-dimension strictly less than and a linear function such that
Thus, if there are too many zoom-outs of a certain co-dimension, then there is necessarily a zoom-out of smaller co-dimension that also works. In that case, the parties could go up to that co-dimension.
This result is not good enough for us, due to the polynomial gap between the agreement between and ’s and and the agreement between an . Indeed, in our range of parameters, will be below the trivial threshold which is the agreement a random linear function has with , and therefore the promise on the function above is meaningless.
We resolve this issue by showing a stronger, essentially optimal version of the above assertion still holds. Formally, we prove:
Lemma 1.7.
For all , there is such that the following holds for . Suppose that is a table that assigns to each subspace of dimension a linear function, and suppose that there are at least subspaces of co-dimension and linear functions such that
for all . Then, there exists a zoom-out of co-dimension strictly smaller than and a linear function such that
We defer a detailed discussion about Lemma 1.7 and its proof to Section 8, but remark that our proof of Lemma 1.7 is very different from the arguments in [DKK+18] and is significantly more involved. Our proof uses tools from [KMS17, DKK+18], tools from the analysis of the classical Plane vs Plane and direct product testing [RS97, IKW12, MZ23], global hypercontractivity [EKL24] as well as Fourier analysis over the Grassmann graph.
1.2.4 The Outer PCP
Our outer PCP game is the outer PCP of [KMS17, DKK+18], which is a smooth parallel repetition of the equation versus variables game of Hastad [Hås01] (or of [KP06] for the application to Quadratic Programming). As in there, we equip this game with the “advice” feature to facilitate zoom-in coordination (as discussed above). For the sake of completeness we elaborate on the construction of the outer PCP below.
We start with an instance of -Lin that has a large gap between the soundness and completeness. Namely, we start with an instance of linear equations over in which each equation has the form . It is known [Hås01] that for all , it is NP-hard to distinguish between the following two cases:
-
1.
YES case: .
-
2.
NO case: .
Given the instance , we construct a -Prover--Round game, known as the smooth equation versus variable game with -advice as follows. The verifier has a smoothness parameter and picks a random equation , say , from . Then:
-
1.
With probability the verifier takes and vectors sampled uniformly and independently.
-
2.
With probability , the verifier sets , chooses a set consisting of a single variable uniformly at random. The verifier picks uniformly and independently and appends to each the value in the coordinates of to get .
After that, the verifier sends and to the first prover and and to the second prover. The verifier expects to get from them assignments to the variables in and in , and accepts if and only if these assignments are consistent, and furthermore the assignment to satisfies the equation .
Denoting the equation versus variable game by , it is easy to see that if , then , and if , then . The gap between and is too weak for us, and thus we apply parallel repetition.
In the parallel repetition of the smooth equation versus variable game with advice, denoted by , the verifier picks equations uniformly and independently , and picks , and , for each from independently. Thus, the questions of the provers may be seen as and and their advice is for and for respectively. The verifier expects to get from the first prover a vector in which specifies an assignment to , and from the second prover a vector in specifying an assignment to . The verifier accepts if and only if these assignments are consistent and the assignment of the first prover satisfies all of . It is clear that if , then . Using the parallel repetition theorem of Rao [Rao11] (albeit not in a completely trivial way) we argue that if , then . The game is our outer PCP game.
Remark 1.8.
We remark that in the case of the Quadratic Programming application, we require a hardness result in which the completeness is very close to in the form of Theorem 2.1. The differences between the reduction in that case and the reduction presented above are mostly minor, and amount to picking the parameters a bit differently. There is one significant difference in the analysis; we require a much sharper form of the “covering property” used in [KMS17, DKK+18], as elaborated on in Section 1.2.6
1.2.5 Composing the Outer PCP and the Inner PCP Game
To compose the outer and inner PCPs, we take the outer PCP game, only keep the questions to the first prover and consider an induced -Prover--Round game on it. The alphabet is , that given a question specifies an assignment to the variables of . There is a constraint between and if there is a question to the second prover such that . Denoting the assignments to and by and , the constraint between and is that satisfies all of the equations that form , satisfies all of the equations that form , and , agree on .
The composition amounts to replacing each question with a copy of our inner PCP. Namely, we identify between the question and the space , and then replace by a copy of the sub-spaces graph of . The answer is naturally identified with the linear function , which is then encoded by the sub-spaces encoding via tables of assignments and .
The constraints of the composed PCP must check that: (1) side conditions: the encoded vector satisfies the equations of , and (2) consistency: and agree on .
The first set of constraints is addressed by the folding technique, which we omit from this discussion. The second set of constraints is addressed by the vs subspace test, except that we have to modify it so that it works across blocks and . This completes the description of the composition step of the other PCP and the inner PCP, and thereby the description of our reduction.
1.2.6 The Covering Property
We end this introductory section by discussing the covering property. The covering property is an important feature of our outer PCP construction which enables the composition step to go through. The covering property first appeared in [KS13] and later more extensively in the context of the -to- Games [KMS17, DKK+18]. To discuss the covering property, let be thought of as large, let be thought of as and consider sets consisting of distinct element, each has size (in our context, will be the set of variables in the th equation the verifier chose). Let , and consider the following two distributions over tuples in :
-
1.
Sample uniformly.
-
2.
For each independently, take with probability and otherwise take randomly of size , then set . Sample uniformly and lift them to points in by appending ’s in . Output the lifted points.
In [KMS17] it is shown that the two distributions above are close in statistical distance, which is good enough for the purposes of Theorem 1.3. However, this is not good enough for Theorem 1.4. 444 The reason is that letting be the size of the instance we produce, it holds that is roughly logarithmic and is the alphabet size. To have small statistical distance, we must have , hence the soundness error could not go lower than . Carrying out a different analysis, we are able to show that the two distributions are close with better parameters and in a stronger sense: there exists a set of tuples which has negligible measure in both distributions, such that each tuple not in is assigned the same probability under the two distribution up to factor . We are able to prove this statement provided that is only slightly larger than .
The issue with the above two distributions is that they are actually far from each other if, say, . To see that, one can notice that the expected number of ’s such that each one of has the form on coordinates corresponding to is very different. In the first distribution, this expectation is which is less than 1, whereas in the second distribution it is at least .
To resolve this issue and to go all the way through in the Quadratic Programming application, we have to modify the distributions in the covering property so that (a) they will be close even if , and (b) we can still use these distributions in the composition step in our analysis of the PCP construction. Indeed, this is the route we take, and the two distributions we use are defined as follows:
-
1.
Sample uniformly.
-
2.
For each independently, take with probability and otherwise take randomly of size , then set . Sample uniformly, and let be the vector that has on coordinates of and everywhere else. Lift the points to by appending ’s in and take where are independent random elements from . Output .
We show that for a suitable choice of and , these distributions are close even in the case that . 555More speifically, one takes a small and chooses , . Indeed, as a sanity check one could count the expected number of appearances of blocks of the form and see they are very close ( versus ). In this setting of parameters, is roughly equal to the alphabet size – which can be made to be equal under quasi-polynomial time reductions – it is sufficient to get the result of Theorem 1.4.
Remark 1.9.
We remark that a tight covering property is crucial for obtaining the tight hardness of approximation factor in Theorem 1.4. In the reduction from -Prover--Round games to Quadratic Programs, which is due to [ABH+05], the size of the resulting instance is exponential in the alphabet size and the soundness error remains roughly the same. In our case the alphabet size is roughly hence the instance size is dominated by . If our analysis required , then even showing an optimal soundness of for the -Prover--Round game would only yield a factor of hardness for quadratic programming.
2 Preliminaries
2.1 The Grassmann Graph
In this section we present the Grassmann graph and some Fourier analytic tools on it that are required for our analysis of the inner PCP. Throughout this section, we fix parameters with , and a prime power .
2.1.1 Basic Definitions
The Grassmann graph is defined as follows.
-
•
The vertex set corresponds to the set of -dimensional subspaces .
-
•
The edge set corresponds to all pairs of -dimensional subspaces such that .
At times we will have a vector space over , and thus we may identify with and work with the Grassmann graph on the -dimensional subspaces . We may also use to denote this graph, which is isomporphic to if . Abusing notation, we also use to denote the set of all -dimensional subspaces in . Throughout, we denote by the set of complex valued functions .
The number of -dimensional subspaces of is counted by the Gaussian binomial coefficient, . The following standard fact gives a formula for the Gaussian binomial coefficients, and we omit the proof.
Fact 2.1.
Suppose , then the number of vertices in is given by
Abusing notations, we denote by the set of dimensional subspaces of .
Zoom ins and Zoom outs.
A feature of the Grassmann graph is that it contains many copies of lower dimensional Grassmann graphs as induced subgraphs. These subgraphs are precisely the zoom-ins and and zoom-outs referred to in the introduction, and they play a large part in the analysis of our inner PCP and final PCP. For subspaces , let
We refer to as a zoom-in and as a zoom-out. When , is the zoom-in on , and when , is the zoom-out on .
2.1.2 Pseudo-randomness over the Grassmann graph
One notion that will be important to us is -pseudo-randomness, which measures how much can deviate from its expectation on a zoom-in/zoom-out restrictions of “size ”. For all of our applications, and will both be indicator functions of some sets of vertices, so it will be helpful to think of this case for the remainder of the section. 666We remark that the results we state have more general versions that apply to wider classes of functions. We refrain from stating them in this generality for sake of simplicity. Let (for indicator functions, this is simply the measure of the indicated set). For subspaces , define
Definition 2.2.
We say that a Boolean function is -pseudo-random if for all satisfying , we have
We will often say that a set is -pseudo-random if its indicator function is. Because the Grassmann graph is not a small-set expanders, there are small sets in it that do not look “random” with respect to some combinatorial counting measures (such as edges between sets, expansion and so on). Intuitively, a small set which is highly pseudo-random will exhibit random-like structure with respect to several combinatorial measures of interest, and the two lemmas below are instantiations of it required in our proof. The proof proceed by reducing them to similar statements about the Bi-linear scheme, which can then be proved directed by appealing to global hypercontractivity results of [EKL23, EKL24].
For the analysis of the inner PCP, we require the following lemma, which bounds the number of edges between and when is -pseudo-random.
Lemma 2.3.
Let and be Boolean functions such that , , and suppose that is pseudo-random. Then for all that are powers of ,
Proof.
Deferred to Section A ∎
We also need the following lemma, asserting that if a not-too-small set is highly pseudo-random, then its density remains nearly the same on all zoom-ins.
Lemma 2.4.
For all , the following holds for sufficiently large . Suppose that , , and let be a subspace such that . Let have measure and set . If is -pseudo-random, then
Proof.
The proof is deferred to Appendix A.4 ∎
At times we will also use the term global to refer to sets whose indicator functions are pseudo-random. That is, we say that a set is -global if its indicator is -pseudo-random.
2.2 Hardness of 3LIN
In this section we cite several hardness of approximation results for the problem of solving linear equations over finite fields, which are the starting point of our reduction. We begin by defining the and the problem.
Definition 2.5.
For a prime power , an instance of is which consists of a set of variables and a set of linear equations over . Each equation in depends on exactly three variables in , each variable appears in at most equations, and any two distinct equations in share at most a single variable.
The goal in the problem is to find an assignment satisfying as many of the equations in as possible. The maximum fraction of equations that can be satisfied is called the value of the instance. We remark that usually in the literature, the condition that two equations in share at most a single variable is not included in the definition of , as well the the bound on the number of occurences of each variable.
For , the problem is the promise problem wherein the input is an instance of promised to either have value at least or at most , and the goal is to distinguish between these two cases. The problem with various settings of and will be the starting point for our reductions.
To prove Theorem 1.3, we shall use the classical result of Håstad [Hås01]. This result says that for general instances (i.e., without the additional condition that two equations share at most a single variable), the problem is NP-hard for all constant and . This result implies the following theorem by elementary reductions:
Theorem 2.1.
There exists such that for every constant and prime , is -hard.
To prove Theorem 1.4 we will need a hardness result for with completeness close to , and we will use a hardness result of Khot and Ponnuswami [KP06]. Once again, their result does not immediately guarantee the fact that any two equations share at most a single variable, however once again this property may be achieved by an elementary reduction.
Theorem 2.2.
There is a reduction from with size to a instance with size over a field of characteristic , where,
-
•
Both and the running time of the reduction are bounded by
-
•
.
-
•
.
3 The Outer PCP
In this section, we describe our outer PCP game. In short, our outer PCP is a smooth parallel repetition of the variable versus equation game with advice. This outer PCP was first considered in [KS13] without the advice feature, and then in [KMS17] with the advice feature.
3.1 The Outer PCP construction
Let be parameters that determine the completeness and the soundness our . Our reduction starts with the problem, and we fix an instance of it for the rest of this section. Our presentation is gradual, and we begin by presenting the basic Variable versus Equation Game. We then equip it with the additional features of smoothness and advice.
3.1.1 The Variable versus Equation Game
We first convert the instance into an instance of -Prover--Round Games, and it will be convenient for us to describe it in the active view with a verifier and provers.
In the Variable versus Equation game, the verifier picks an equation uniformly at random, and then chooses a random variable . The verifier sends the question , i.e. the three variables appearing in , to the first prover, and sends the variable to the second prover. The provers are expected to answer with assignments to their received variables, and the verifier accepts if and only if the two assignments agree on and the first prover’s assignment satisfies the equation . If the verifier accepts then we also say that the provers pass. This game has the following completeness and soundness, which are both easy to see (we omit the formal proof):
-
1.
Completeness: If has an assignment satisfying -fraction of the equations, then the prover’s have a strategy that passes with probability at least .
-
2.
Soundness: If has no assignment satisfying more than -fraction of the equations, then the prover’s can pass with probability at most .
3.1.2 The Smooth Equation versus Variable Game
We next describe a smooth version of the Variable versus Equation game. In this game, the verifier has a parameter , and it proceeds as follows:
-
1.
The verifier chooses an equation uniformly, and lets be the set of variables in .
-
2.
With probability , the prover chooses . With probability , the prover chooses randomly of size .
-
3.
The verifier sends to the first prover, and to the second prover.
-
4.
The provers respond with assignments to the variables they receive, and the verifier accepts if and only if their assignments agree on and the assignment to satisfies the equation .
The smooth Variable versus Equation game has the following completeness and soundness property, which are again easily seen to hold (we omit the formal proof).
-
1.
Completeness: If has an assignment satisfying fraction of the equations, then the provers have a strategy that passes with probability at least .
-
2.
Soundness: If has no assignment satisfying more than fraction of the equations, then the provers can pass with probability at most .
3.1.3 The Smooth Equation versus Variable Game with Advice
Next, we introduce the feature of advice into the smooth Variable versus Equation Game. This “advice” acts as shared randomness which may help the provers in their strategy; we show though that it does not considerably change the soundness. The game is denoted by for and , and proceeds as follows:
-
1.
The verifier chooses an equation uniformly, and lets be the set of variables in .
-
2.
With probability , the verifier chooses . With probability , the verifier chooses randomly of size .
-
3.
The verifier picks vectors uniformly and independently. If the verifier takes for all , and otherwise the verifier takes the vectors where for all , the vector agrees with on the coordinate of , and is in the coordinates of .
-
4.
The verifier sends and to the first prover, and and to the second prover.
-
5.
The provers respond with assignments to the variables they receive, and the verifier accepts if and only if their assignments agree on and the assignment to satisfies the equation .
Below we state the completeness and soundness of this game:
-
1.
Completeness: If has an assignment satisfying fraction of the equations, then the provers have a strategy that passes with probability at least . This is easy to see.
-
2.
Soundness: If has no assignment satisfying more than fraction of the equations, then the provers can pass with probability at most . Indeed, suppose that the provers can win the game with probability at least . Note that with probability at least it holds that and all the vectors and are all , in which case the provers play the standard equation versus variable game. Thus, the provers’ strategy wins in the latter game with probability at least , and contradiction.
3.1.4 Parallel Repetition of the Smooth Equation versus Variable Game with Advice
Finally, our Outer PCP is then the -fold parallel repetition of , which we denote by . Below is a full description of it:
-
1.
The verifier chooses equations uniformly and independently, and lets be the set of variables in .
-
2.
For each independently, with probability , the verifier chooses . With probability , the verifier chooses randomly of size .
-
3.
For each independently, the verifier picks a vectors uniformly and independently. If the verifier takes for , and otherwise the verifier takes the vectors where for all , the vector agrees with on the coordinate of , and is in the coordinates of .
-
4.
The verifier sets and for each , and and for each . The verifier sends and to the first prover, and and to the second prover.
-
5.
The provers respond with assignments to the variables they receive, and the verifier accepts if and only if their assignments agree on and the assignment to satisfies the equations .
Next, we state the completeness and the soundness of the game , and we begin with its completeness.
Claim 3.1.
If has an assignment satisfying at least of the equations, then the provers can win with probability at least .
Proof.
Let be an assignment that satisfies at least fraction of the equations in , and consider the strategy of the provers that assigns their variables according to . Note that whenver each one of the equations the verifier chose is satisfied by , the verifier accepts. By the union bound, the probability this happens is at least . ∎
Next, we establish the soundness of the game .
Claim 3.2.
If there is no assignment to satisfying at least of the equations, then the provers can win with probability at most .
Proof.
We appeal to the parallel repetition theorem for projection games of Rao [Rao11], but we have to do so carefully. That theorem states that if is a -Prover--Round game with , then . We cannot apply the theorem directly on (as the square is too costly for us). Instead, we consider the game and note that it has value bounded away from .
Write . Note that probability at least there exists at least a single coordinate in which and all of the advice vectors and are all . Thus, there exists a coordinate and a fixing for the questions of the provers outside so that the answers of the players to the th coordinate win the standard equation versus variable game with probability at least . It follows that , and so .
We conclude from Rao’s parallel repetition theorem that
Viewing the advice as subspaces.
Due to the fact that each variable appears in at most equations, it can easily be seen that with probability , all variables in are distinct. In that case, note that the vectors of advice to the second prover, , are uniform, and the second prover may consider their span . Note that the distribution of is that of a uniform dimensional subspace of . As for the second prover, the vectors are not uniformly distributed. Nevertheless, as shown by the covering property from [KS13, KMS17] (and presented below), the distribution of is close to uniform over -tuple of vectors from . Thus, the first prover can also take their span, call it , and think of it as a random -dimensional subspace of (which is highly correlated to ).
4 The Composed PCP Construction
In this section we describe the final PCP construction, which is a composition of the outer PCP from Section 3 with the inner PCP based on the Grassmann consistency test.
4.1 The Underlying Graph
Our instance of -Prover--Round Games starts from an instance of . Consider the game from Section 3, and let denote the set of questions asked to the first prover. Thus consists of all -tuples of equations from the instance . For let denote the indicator vector on the three variables appearing in .
It will be convenient to only keep the that satisfy the following properties:
-
•
The equations are distinct and do not share variables.
-
•
For any and pair of variables and , the variables and do not appear together in any equation in the instance .
The fraction of that do not satisfy the above is which is negligible for us, and dropping them will only reduce our completeness by . This will not affect our analysis, and henceforth we will assume that all satisfy the above properties. We now describe the -Prover--Round Games instance . All vertices in the underlying graph will correspond to subspaces of .
4.1.1 The Vertices
For each question , let , where is the vector with ones at coordinates corresponding to variables appearing in . We can think of the ’s as vectors from an underlying space . By the first property described above, and . The vertices of are:
In words, the vertices on the side are all -dimensional subspaces of some for some . For technical reasons, we require them to intersect trivially (which is the case for a typical -dimensional space) and add to them the space .777This has the effect of collapsing and such that to a single vertex. The vertices on the side are all dimensional subspaces of .
4.1.2 The Alphabets
The alphabets have sizes and . For each vertex , let denote the function that satisfies the side conditions given by the equations in . Namely, if is the equation for , then . We say a linear function satisfies the side conditions of if . In this language, for a vertex we identify with
As and , it is easy to see that the above set indeed has size . For each right vertex , we identify with
4.1.3 The Edges
To define the edges, we first need the following relation on the vertices in . Say that if
Recall that all subspaces above are in hence the direct sums and equality above are well defined. The relation described is in fact an equivalence relation and thus partitions the vertices in into disjoint equivalence classes. It is clear that the relation is reflexive and symmetric, so we need only show that it is also transitive.
Lemma 4.1.
If , and , then
Proof.
We “add” to the second equation to obtain,
Next, write , where is the span of all vectors for equations in that are also in or , while is the span of all vectors for equations that are in neither nor . It follows that . Now note that any equation has at most one variable that appears in an equation in , and at most one variable that appears in an equation in . Thus, each , has a “private variable”, and as the equations in are over disjoint sets of variables, this private variable does not appear in . It follows that
Indeed, by the above discussion any nonzero vector in is nonzero on at least one coordinate of (corresponding to a private variable), and no vector in or is supported on this coordinate.
Substituting into the original equation yields,
Since , this equivalent to
As , it follows that
as desired. ∎
By Lemma 4.1 the relation is indeed an equivalence relation and we may partition into equivalence classes, . We call each class a clique and partition into cliques:
The actual number of cliques, , will not be important, but it is clear that such a number exists. The edges of our graph will be between vertices and if there exists such that . The edges will be weighted according to a sampling process that we describe in the next section, which also explains the constraints on . For future reference, the following lemma will be helpful in defining the constraints:
Lemma 4.2.
Suppose and that is a linear function satisfying the side conditions. Then there is a unique linear function that satisfies the side conditions such that there exists a linear function satisfying the side conditions (of both and ) such that
In words, is a linear extension of both and .
Proof.
Note that there is only one way to extend to in a manner that satisfies the side conditions given by . Let this function be . We take to be . ∎
4.1.4 The Constraints
Suppose that is an assignment to that assigns, to each vertex , a linear function satisfying the side conditions. Further suppose that is an assignment that assigns to each vertex a linear function on . The verifier performs the following test, which also describes the constraints of :
-
1.
Choose uniformly at random from .
-
2.
Choose uniformly, where and , and choose of dimension uniformly.
-
3.
Choose uniformly
-
4.
As in Lemma 4.2, extend to in the unique manner that respects the side conditions and let be the restriction of this extension to .
-
5.
Accept if and only if .
This finishes the description of our instance . It is clear that the running time and instance size is and that the alphabet size is .
Before arguing about the completeness and soundness, we will present some necessary tools. As is usually the case, showing completeness is relatively easy, and all of the tools presented are for the much more complex soundness analysis.
5 Tools for Soundness Analysis
In this section we will present all of the tools needed to analyze the soundness of our PCP.
5.1 The versus subspace agreement test
We begin by discussing the versus test and our decoding theorem for it. In our setting, we have a question for the first prover, and we consider the versus test inside the space . In our setting this test passes with probability at least (where is, say ) and we will want to use this fact to devise a strategy for the first prover. Below, we first state and prove a basic decoding theorem, and then deduce from it a quantitative better version that also incorporates the side conditions.
Let be a table that assigns, to each , a linear function , and let be a table assigning to each a linear function . We recall that . In this section, we show that if tables and are -consistent, namely
for , then the table must have non-trivial agreement with a linear function on some zoom-in and zoom-out combination of constant dimension. To prove that, we use Lemma 2.3 along with an idea from [BKS19].
Theorem 5.1.
Suppose that tables and are -consistent where . Then there exist subspaces and a linear function such that:
-
1.
.
-
2.
for -fraction of -dimensional ,
where .
Proof of Theorem 5.1.
Consider the bipartite graph whose sides are the vertices of and , and its set of edges consists of pairs such that . Consider the normalized adjacency operator of , and let be its adjoint operator.
Choose a linear function uniformly at random and define the (random) sets of vertices
Denote by the set of edges with endpoints in and . We lower bound the expected size of over the choice of . Note that for each edge such that , we have that with probability . Indeed, with probability we have that , and in that case we automatically get that . As the number of edges such that is at least , we conclude that
Note that we also have that
Using Linearity of Expectation, we get that
thus there exists for which the random variable on the left hand side is at least , and we fix so that
(2) |
We claim that is not -pseudo-random for and . Suppose for the sake of contradiction that this is not the case, and that is -pseudo-random. Denote and . By Lemma 2.3 for any that is a power of we have
(3) |
In the last inequality, we used the fact that by (2)
so , and thus the second term on the middle of (3) is negligible compared to the first term there. Combining (2) and (3) gives us that
Simplifying, using the definition of , the fact that and the fact that we get
Investigating the second two exponents of , we have that for ,
This implies that
and contradiction. It follows that is not -pseudo-random, and unraveling the definition of not being pseudo-random gives the conclusion of the theorem. ∎
5.1.1 Finding a Large Fraction of Successful Zoom-Ins
Theorem 5.1 asserts the existence of a good pair of zoom-in and zoom-out on which the table has good agreement with a global linear function. As discussed in the introduction, our argument requires a quantitatively version asserting that there is a good fraction of zoom-ins that work for us. Below, we state a strengthening of Theorem 5.1 achieving this; it easily follows from Theorem 5.1, and we defer the proof to Section B.
Theorem 5.2.
Suppose that tables and are -consistent for . Then there exist positive integers and satisfying , such that for at least -fraction of the -dimensional subspaces , there exists a subspace of codimension and a linear function such that
Proof.
The proof is deferred to Section B. ∎
5.1.2 Incorporating Side Conditions for Zoom-Ins
Next, we require a version of Theorem 5.2 which also takes the side conditions into account.
Theorem 5.3.
Let be a question to the first prover, let the first prover’s table, including the side conditions, and suppose that
Then there are parameters and such that , such that for at least fraction of the -dimensional subspaces , there exists of codimension containing , and a global linear function that respects the side conditions on such that
Proof.
For any -dimensional subspace such that , let be the table given by for all -dimensional subspaces . We can choose a -dimensional subspace such that by first uniformly choosing such that , and then choosing of dimension uniformly. Thus, defining
we have
In particular, for at least -fraction of ’s, we have . For such ’s, by Theorem 5.2, there exist positive integers and such that for at least -fraction of -dimensional zoom-ins , there exists a zoom-out of co-dimension and a linear function such that,
Let and let be the unique extension of to satisfying the side conditions. We claim that
Indeed, for each there are an equal number of such that and , so
To conclude, we see that sampling and then of dimension , we get that there is a zoom-out and a function satisfying the conditions in the theorem with probability at least . As the marginal distribution over is -close to uniform over all -dimensional subspaces the conclusion follows. ∎
5.2 The Covering Property
In this section, we present the so called “covering property”, which is a feature of our PCP construction that allows us to move between the first prover’s distribution over -dimensional subspaces of and the second prover’s distribution over -dimensional subspaces of . Similar covering properties are shown in [KS13, KMS17]; however, obtaining the optimal quadratic-programming hardness result in Theorem 1.4 requires a stronger analysis that goes beyond the covering properties of [KS13, KMS17]. We are able to obtain a covering property with the following parameters:
(4) |
where is a constant arbitrarily small relative to .
5.2.1 The Basic Covering Property
To start, we state a basic form of the improved covering property that is required in our analysis and defer its proof to Appendix C. Fix a question to the first prover and recall that where is the vector that is one at coordinates corresponding to variables in and elsewhere. The covering property we show will relate the following two distributions:
-
•
Choose uniformly.
-
•
Output the list .
-
•
Choose according to the Outer PCP.
-
•
Choose uniformly, and lift these vectors to by inserting ’s into the missing coordinates.
-
•
Choose uniformly, and set for .
-
•
Output the list .
With these two definitions, the covering property used in prior works asserted that the distribution is statistically close to a variant of the distribution . This closeness is not good enough for us, as we will want to consider events of rather small probability under and still assert that their probability is roughly the same in . First, in these earlier works, the distribution was generated by a similar process to the above without the addition of the random vectors from . As explained in the introduction however, this distribution is not good enough for the purpose of Theorem 1.4, and we must consider the distribution above. Second, the notion of statistical closeness is too rough for us, and we show that in fact, almost all inputs are assigned the same probability under these two distributions up to factor .
More precisely, set throughout this subsection. Our covering property is the following statement:
Lemma 5.4.
Let be a parameter such that . There exists a small set such that both and are at most , and for all we have
Proof.
The proof is deferred to Appendix C.1. ∎
5.2.2 The Covering Property with Zoom-ins
Lemma 5.4 represents the most basic form of the covering property, and for out application we require a version of it that incorporates zoom-ins and advice. Namely, we will actually interested in the case where and are conditioned on some -dimensional zoom-in , for an arbitrary dimension . To make notation simpler, let us write and use to denote . Additionally, define
From Lemma 5.4 we can conclude that for any that is not too small, the measure is within at least a constant factor of for nearly all .
Lemma 5.5.
For any , we have
where is the span of uniformly random vectors in .
Proof.
Throughout the proof all of the expectations and probabilities over choose as in the lemma statement. Let be the small set of points from Lemma 5.4. By assumption we have
Thus, by Markov’s inequality, we have that with probability at least , we have . In this case we have,
where we applied Lemma 5.4 in the second transition. Dividing both sides by gives that
(5) |
It follows that
where we apply Equation (5) in the second transition and the assumption in the last transition. ∎
5.2.3 The Covering Property for the Advice
We will also need a similar, and simpler, version of the above lemma that applies to -dimensional subspaces for some constant . This is to handle the fact that the zoom-in is sampled uniformly from after is chosen according to the outer PCP, and then lifted to a subspace over , instead of uniformly from . Formally, let denote the former distribution over -dimensional subspaces and let denote the latter distribution over -dimensional subspaces . These are the same as the distributions and of the previous subsection except over instead of . We show the following.
Lemma 5.6.
Let be a set of -dimensional subspaces in satisfying . Then,
5.3 The Number of Maximal Zoom-Outs is Bounded
In Theorem 5.2, we showed that the two provers can agree on a zoom-in with reasonable probability using their advice. The same cannot be said for zoom-outs however, and to circumvent this issue we must develop further tools. In this section, we define the notion of maximal zoom-outs and show that for a fixed zoom-in , the number of maximal zoom-outs is bounded.
5.3.1 Generic Sets of Subspaces
One of our primary concerns with respect to zoom-outs is that it is possible for a prover to have many good zoom-outs to choose from (so that independent sampling doesn’t work) but not enough to allow for advice-type solution. To deal with large collections of zoom-outs we define a special property of zoom-outs that is called “genericness”. To motivate it, note that if are distinct subspaces of co-dimension , then is a subspace whose co-dimension is between and . For a typical pair of subspaces, the intersection has dimension , in which case we say they are generic. Genericness is useful probabilistically, since if are generic, then the event that a randomly chosen -dimensional subspace is contained in , and the event it is contained in , are almost independent. Below is a more general and formal definition:
Definition 5.7.
We say that a set of codimension subspaces of is -generic with respect to if for any -distinct subspaces, say , we have . When the ambient space is clear from context we simply say that is -generic.
We remark that any set of subspaces that is -generic with respect to is also -generic with respect to for any . In this section, we will show a couple of results regarding generic sets of subspaces that will be used to bound the number of maximal zoom-outs in Section 8.
The result we need is a sunflower-type lemma, stating that any large set of codimension subspaces inside contains a large set of subspaces that are -generic with respect to for some . Below is a formal statement.
Lemma 5.8.
Let be integers and let be a set of subspaces of co-dimension inside of . Then there exists a subspace and a set of subspaces such that:
-
•
.
-
•
Each is contained in and has co-dimension with respect to , where .
-
•
is -generic with respect to .
In order to show Lemma 5.8, we introduce two necessary lemmas. The first, Lemma 5.9, states that for , any -generic set of subspaces contains a large -generic set of subspaces. The second, Lemma 5.10,states that either a set of subspaces is already -generic, or there are many subspaces in the set that are contained in the same hyperplane. Using this lemma, we can start from a large set of subspaces inside of an ambient space and iteratively reduce to the dimension of the ambient space until we find a -generic set of subspaces relative to the ambient space. Indeed, either the set is already -generic, or there is a hyperplane such that the set of subspaces in contained in is large. Taking this set to be the new and to be the new ambient space, we obtain a, still, large set of subspaces, whose codimension is now one less. We may repeat this process until is a set of hyperplanes in the ambient space, at which point it will be -generic.
Lemma 5.9.
Let be a set of -subspaces of codimension inside of that is -generic with respect to , then there is a subset of size that is -generic with respect to .
Proof.
Fix any distinct subspaces in , say and let . Since is -generic, . We claim that there are at most subspaces such that . Call such subspaces bad and suppose for the sake of contradiction that there are greater than bad subspaces . Then for each bad we have,
Therefore, for each , the space is contained in a hyperplane such that . There are at most hyperplanes containing , and by the pigenhole principle it follows that there are two bad subspaces say that are both contained in the same hyperplane . This is a contradiction however, as by the -genericness of , we must have
and hence and cannot both be contained in the hyperplane .
The lemma now follows from the claim we have just shown. Construct a subset greedily as follows:
-
1.
Initialize by picking arbitrary subspaces from and inserting them to .
-
2.
For any subspaces in , say , remove any which is bad for them.
-
3.
If is not empty, pick some , insert it to and iterate.
Note that trivially, the collection will be -generic in the end of the process. To lower bound the size of , note that when , the number of elements from that have been deleted is at most , and hence so long as this value is at most , we may do another iteration. Thus, we must have that when the process terminates. ∎
Lemma 5.10.
Let be a set of subspaces of of codimension . Then for any integer , at least one of the following holds.
-
•
There are subspaces, way such that for every pair , .
-
•
There is a subspace of co-dimension that contains of these subspaces, say .
Proof.
Note that for any , we have . Consider the graph with vertices with an edge if and only if and . If every vertex in this graph has degree at most , then we are done as there is an independent set of size and these subspaces satisfy the first condition. Suppose this is not the case. Then there is a vertex, say , that has neighbors, say . For , we have , so
Thus is always contained in a codimension subspace of that contains . Since the number of such subspaces is , there must exist one subspace, say , that contains at least of the subspaces in the list . ∎
Repeatedly applying Lemma 5.10 yields the following corollary.
Corollary 5.11.
Let be a set of subspaces of of codimension with respect to . There exists a subspace , an integer , and a subset of , say , all contained in such that,
-
•
Each , , has codimension with respect to .
-
•
Each , , has codimension .
Proof.
To start set . If the ’s have codimension in then the result holds.
Otherwise, if the conclusion does not hold, then apply Lemma 5.10 with . Either the first condition of Lemma 5.10 holds and we are done, or we can find a new subspace, , of codimension inside the current containing at least of the ’s. Set and repeat. Note that the codimension of the ’s with respect to drops by after every iteration, so we will repeat at most times before reaching the desired conclusion. This yields a list of ’s that satisfy the conditions of size at least
Proof of Lemma 5.8.
In addition to Lemma 5.8, we state another useful feature of generic sets of subspaces, formalized in Lemma 5.12 below. The lemma asserts that if a collection is generic, and one zooms-outs from the ambient space into a hyperplane , then one gets an induced collection which is almost as generic.
Lemma 5.12.
Let be a set of subspaces of codimension that is -generic with respect to some space for an even integer , and let be a hyperplane in . Then the set of subspaces can be made a -generic set of subspaces with respect to with codimension inside of by removing at most subspaces from it.
Proof.
Suppose that is not -generic with respect to with codimension inside of , as otherwise we are done. In this case, there must exist distinct subspaces, say such that
where the codimension is with respect to . However, since is -generic (and thus -generic as well) with respect to , this implies that
Now delete from . We claim that the resulting set is -generic with respect to . Suppose for the sake of contradiction that it is not. Then there must be another distinct subspaces, say such that
This would imply
This is a contradiction however, as is -generic with respect to , so , and
and contradiction. ∎
Lemma 5.13.
Let be a set of subspaces that is generic with respect to and let be a subspace of codimension . Then, the set of subspaces,
can be made generic with respect to by removing at most subspaces.
Proof.
There is a sequence of subspaces , such that is a hyperplane inside of . Do the following,
-
1.
Initialize and set .
-
2.
Set , and then remove the minimal number of subspaces to turn into a -generic collection with respect to .
-
3.
Stop if , otherwise, increase by and return to step .
It is clear that the output is a set of subspaces that is -generic with respect to . Furthermore, during each iteration, at most subspaces are removed by Lemma 5.12, and the result follows. ∎
5.3.2 The Sampling Lemma
As explained earlier, the notion of genericness is useful probabilistically, and in this section we state and prove a sampling lemma about generic collections which is necessary for our analysis. Fix an arbitrary zoom-in of dimension , and let be a -generic collection of subspaces of of codimension all containing . Also let be a set of -dimensional subspaces containing . For the remainder of this subsection, use to denote the set of -dimensional subspaces in containing . Consider the following two probability measures over :
-
1.
The distribution which is uniform over .
-
2.
The distribution , wherein a subspace is sampled by first picking uniformly and then sampling a subspace from uniformly.
The main content of this section is the following lemma, asserting that the measures and are close in statistical distance provided that is large. More precisely:
Lemma 5.14.
For any we have
We now set up some notations for the proof of Lemma 5.14. For let
and for an arbitrary pair of distinct define the following quantities:
(6) |
We note that all of these quantities are well defined as they do not depend on the identity of and . The first two equations are clear and the third uses the fact that is -generic. Thus,
(7) |
For the first equation, we are using the fact that , while the second equation is evident. In the following claim we analyze the expectation and variance of when is chosen uniformly form :
Claim 5.15.
and . where the variance is over uniform .
Proof.
By linearity of expectation
and we move on to the variance analysis. To bound , write
It follows that,
Finally note that and are nearly the same value, and it can be checked using Equation (6) that and so . ∎
Combining Chebyshev’s inequality with Claim 5.15, we conclude the following lemma which will be useful for us later on.
Lemma 5.16.
For any it holds that
Proof.
This is an immediate result of Chebyshev’s inequality with the bounds from Claim 5.15. ∎
5.3.3 Bound on Maximal Zoom-outs
For this subsection, we work in the second prover’s space, , and make the assumption that , say to be concrete. We first establish several results in the simplified setting where there is no zoom-in. After that we show how to deduce an analogous result with a zoom-in. Throughout this section, we fix to be a table that assigns to each a linear function on .
Definition 5.17.
Given a table on and a subspace , we call a zoom-out, function pair, , where and , -maximal with respect to on if
and there does not exist another zoom-out function pair, such that , , and
In the case that , we say that is -maximal with respect to .
In the above statement, should be thought of as small and should be thought of as an absolute constant. With this in mind, a zoom-out and a linear function on it is called maximal if there is no zoom-out that strictly contains , and an extension of to , that has the same agreement with as (up to constant factors). As an immediate consequence of the definition of -maximal, we have the following lemma, which roughly states that every zoom-out for which there that has a linear function with good agreement inside the zoom-out is contained in a maximal zoom-out (with only slightly worse agreement).
Lemma 5.18.
Let be a table on , , and be a subspace of codimension containing . Suppose that there exists a linear function such that
Then there exists a subspace and a linear function such that and is -maximal and a linear function such that
Proof.
This is an immediate consequence of Definition 5.17. If is -maximal then we are done. Otherwise, there must exist such that , and
We can repeat this argument at most times before obtaining some that is maximal and satisfies and . ∎
The following result, which is key to our analysis gives an upper bound on the number of maximal zoom-outs. This lemma is in fact equivalent to Lemma 1.7, and its proof is deferred to Section 8. In order to present the lemma cleanly, we set the following parameters for the remainder of the section, which can all be considered constant
Lemma 5.19.
Let be a table on with and set , , and
Suppose that are zoom-out, function pairs such that the ’s are all distinct and for each , , and
Then there is a a subspace , a linear function , and a set of subspaces of size
-
•
Each is strictly contained in and has codimension with respect to .
-
•
is -generic with respect to .
-
•
For any , .
Proof.
The proof is deferred to Section 8. ∎
To bound the number of maximal zoom-outs, we will also need the following list decoding property.
Lemma 5.20.
Let be a table on , let be an -dimensional subspace, and let be a subspace of codimension . Suppose that is sufficiently large and . Let be a list of distinct linear functions such that for at least -fraction of the -dimensional subspaces such that , for , and . Then,
Proof.
The proof is deferred to Appendix D ∎
Combining Lemma 5.19 with Lemma 5.20, yields a bound on the number of maximal zoom-out function pairs with respect to a table on .
Theorem 5.21.
For any table on such that , any subspace of dimension and any , the number of -maximal zoom-out, function pairs with respect to on is at most .
Proof.
Suppose for the sake of contradiction that are distinct pairs that are maximal with respect to on . By Lemma 5.20, for each , there are at most functions such that for at least -fraction of the . Thus, there are distinct ’s appearing in the pairs, and there is a codimension such there are pairs, say, zoom-out function pairs that are maximal with respect to on such that the ’s are all distinct and of codimension in .
Write . For each of dimension containing , there is a unique such that . Define the table that assigns linear functions to each by
(8) |
For each , let be the unique subspace such that . We have that for at least -fraction of .
By Lemma 5.19 there exists a subspace , a linear function , and a subcollection of size at least such that
-
•
Each is has codimension with respect to .
-
•
is -generic with respect to .
-
•
For any , .
Let and extend to the function on so that, for at least of the in . It follows that there is a set of size .
Furthermore, because is -generic inside of , is -generic inside of . We will now finish the proof by applying Lemma 5.14 on . Specifically, let denote the measure over generated by choosing and then conditioned on , let denote the uniform measure over , and let
Since for every , we have
By Lemma 5.14 with and , it follows that
Summing everything up, this shows that there is a zoom-out function pair such that and for at least one , and . This contradicts the assumption that is -maximal with respect to on . ∎
5.4 An Auxiliary Lemma
We conclude this section with an auxiliary lemma that will be used in the analysis.
Lemma 5.22.
Let be a fixed question to the first prover in the Outer PCP consisting of -variables in some set of equations. Let be a random question to the second prover chosen according to the Outer PCP. Let be a subspace of co-dimension . Then, with probability at least over the choice of the question , we have
Proof.
Say that is given by the constraints for linearly independent. We can view as being defined by the constraints , where correspond to the variables in . The event is equivalent to the event that are linearly independent. Since are linearly independent, there are -coordinates such that the restrictions of to these -coordinates are linearly independent. If none of are in this set of coordinates, then are linearly independent. Conditioned on the size of being , this event happens with probability at least , and as the expectation of is it follows that the probability in question is at least . ∎
6 Analysis of the PCP
In this section we show completeness and soundness analysis of the composed PCP construction from Section 4. As usual, the completeness analysis is straightforward and the soundness analysis will consist the bulk of our effort.
6.1 Completeness
Suppose that the instance has an assignment that satisfies at least of the equations in . Let be the set of all where all equations are satisfied. Then, . We identify with the linear function from , assigning the value to the th elementary basis element . Abusing notation, we denote this linear map by as well.
For each and vertex , we set . Since , these assignments satisfy the side conditions. For all other ’s, set so that the side conditions of are satisfied and . Such an assignment is possible because . Similarly, the table is defined as .
Sampling a constraint, note that the constraint is satisfied whenever the chosen in step of the test satisfies that . As the marginal distribution of is uniform, 888This is true because first a clique is chosen with probability that is proportional to its size and then a vertex is sampled uniformly from the clique. the distribution of is uniform. It follows that the constraint is satisfied whenever , which happens with probability at least . Thus, .
6.2 Soundness
In this section we relate the soundness of the composed PCP to that of the outer PCP and prove Lemma 6.1. More precisely, we show:
Lemma 6.1.
The rest of Section 6 is devoted to the proof of Lemma 6.1. The proof heavily relies on the tools from Section 5. Assume, as in lemma Lemma 6.1, that the , and suppose for the sake of contradiction that there are tables and that are -consistent for . To arrive at a contradiction, we show that this implies strategies for the two provers that with success probability greater than .
6.2.1 Clique Consistency
To start, we will reduce to the case where satisfies a condition called clique-consistency.
Definition 6.2.
We say an assignment to is clique consistent if for every vertex and for every , the assignments and satisfy the -to- constraint between and as specified in Lemma 4.2.
The following lemma shows that if and are -consistent assignments to , then there are clique-consistent assignments and that are also -consistent.
Lemma 6.3.
Suppose that the assignments and are -consistent, then there is a clique-consistent assignment such that and are -consistent.
Proof.
Partition into cliques, . For each , choose a random uniformly, and for every assign in the unique way that is consistent with and the side conditions of as described in Lemma 4.2. It is clear that is clique consistent, and we next analyze the expected fraction of constraints that and satisfy.
Note that an alternative description of sampling a constraint in proceeds as follows. First choose a clique with probability that is proportional to by its size, and then choose in the first step. The rest of the sampling procedure is the same. Let be the probability that the test passes conditioned on being chosen in the second step. It is clear that every vertex in the clique has equal probability of being chosen, therefore the probability of passing if chosen is
On the other hand, the expected fraction of constraints satisfied by and (over the randomness of choosing ) is
To see this, note that for any , is the probability that is used to define on . If this is the case, then the probability the test passes on within is .
Since this holds over every clique, it follows that the expected fraction of constraints satisfied by equals the fraction of constraints satisfied by and . In particular, there is a choice of such that together with it satisfies at least fraction of the constraints. ∎
Applying Lemma 6.3 we conclude that there are clique-consistent assignments to that are -consistent, and henceforth we assume that are clique-consistent to begin with. We remark that, in the notation of Section 4.1.4, the benefit of having a clique-consistent assignment is that the constraint that the verifier checks is equivalent to checking that . The latter check is a test which in performed within the space of the first prover. We will use this fact in the next section.
6.2.2 A Strategy for the First Prover
Let be the consistency of and conditioned on being the question to the first prover. As we are assuming that the overall success probability is at least , . By an averaging argument, for at least -fraction of the ’s. Call such ’s good and let be the set of good ’s.
Let be the question to the first prover and let be the advice. If , then the first prover gives up, so henceforth assume that . For such , the test of the inner PCP passes with probability at least . More concretely, we have
Next, the first prover chooses an integer uniformly, and takes to be the span of the first -advice vectors. By Theorem 5.3, there are satisfying such that for at least of the of dimension , there exists containing of codimension and a linear function that satisfies
(9) |
For simplicity, set . With probability at least , they choose , where is the parameter from Theorem 5.3. Call these dimension subspaces lucky and let be the set of all lucky . For our analysis, we only analyze the case where the first prover chooses chooses , which occurs with probability at least
For each such that and , the first prover chooses a of codimension at most and linear function that satisfies the side conditions on and Equation (9). For such that are in and satisfy , define
For such that or define . Finally, define
and let denote set of -dimensional subspaces such that
(10) |
where . By definition of and , if , we have
where the second transition is because, by definition, every is contained in . For the third transition, we are ignoring the probability that the ’s are linearly dependent and . Indeed, the probability that either of these events occur is at most , and is negligible anyways.
Continuing, for , we have
(11) |
where in the third transition uses Equation (9) to lower bound the first term by .
If either , , or , then the first prover gives up. Otherwise the prover extends the function to a linear function on the entire space randomly, and we denote this extension by . The prover outputs the string as their answer where is the unique string such that for all . As , and by extension , respects the side conditions, it follows that satisfies the -linear equations of .
6.2.3 A Strategy for the Second Prover
Let be the question to the second prover. The second prover will use a table to derive their strategy. The table is obtained from as follows. For a question to the second prover, let be an arbitrary question to the first prover. For all -dimensional subspaces , define
In order to make sure that is well defined, we note two things. First, the subspace can be viewed as a subspace of because each can be “lifted” to a subspace of by inserting ’s into the coordinates corresponding to . Second, note that the choice of does not matter when defining . Indeed, for a fixed , the vertices over all are in the same clique. Since is clique consistent, it does not matter which is chosen when defining , as all choices lead to the same function . Therefore the second prover can construct the table .
After constructing , the second prover then chooses a dimension uniformly for the advice . Note that with probability at least the second prover also chooses . The second prover then chooses a zoom-out function pair that is
-maximal with respect to on if one exists (and gives up otherwise).
Finally, the second prover extends the function randomly to a linear function on to arrive at their answer. The resulting function is linear and it is equal to the inner product function for some unique string . The second prover outputs as their answer.
6.2.4 The Success Probability of the Provers
In order to be successful, a series of events must occur. We go through each one and state the probability that each occurs. At the end this yields a lower bound on the provers’ success probability. We remark that the analysis of this sections requires Lemmas 5.4 and 5.6, so recall that and are set according to Equation (4) in Section 1.2.6 so that these lemmas hold.
First, the provers need , which occurs with probability at least . Assuming that this occurs, the provers then both need to choose for the dimension of their zoom-in, which happens with probability at least . If both provers choose , they both receive advice as the span of random vectors.
The provers then need , , and . When analyzing the probability that these three events occur, we need to recall that the advice vectors are actually drawn according to distribution , the distribution described in Section 5.2.3. We will analyze the probability that the three events occur under and then appeal to the covering property of Lemma 5.6. By Theorem 5.2, the first item occurs with probability at least . On the other hand the probability that the second item does not occur is at most , while the probability that the third item does not occur is at most by Lemma 5.5. Altogether we get that with probability at least
under , we have , , and . By Lemma 5.6, we have that , , and with probability at least under - the distribution which the is actually drawn from.
Now let us assume that , and both provers receive an -dimensional advice such that , , and . The first prover chooses the function . Write . Since , by Equation (6.2.2) we have
since by definition contains at least -fraction of , which is in turn at least fraction of . Next because , we have
(12) |
by Equation (10).
Now let . By Lemma 5.22, with probability at least we have that inside of . Combining this with an averaging argument on Equation (12), we have that with probability at least over ,
(13) |
and . We call such consistent. In the probability above, and henceforth, we view vectors as vectors in with ’s appended in the missing coordinates. At this point there is the slight issue that does not actually correspond to a random entry of the second prover’s table, . Indeed, the second prover can only choose , choose some question to the first prover that contains , lift these to by inserting zeros in the missing coordinates, and look at the entry where . They do not know the question and the side conditions , and hence could not sample the . However, notice that for any , we have that,
We can thus view the being sampled and added to as a virtual step. In the next two equations, we ignore the probability that or are not linearly independent or have spans intersecting to make the expressions above simpler. This probability is at most and is negligible anyways. We have,
This last probability can be related to
By Lemma 5.18, there exists some that is -maximal and satisfies , is linear, and . By Theorem 5.21, the number of is at most
where is some function depending only on . Thus, the second prover chooses with probability at least . Finally, if the second prover chooses , then the provers succeed if both provers extend their functions, and in the same manner. This occurs with probability at least .
Putting everything together, we get that the provers succeed with probability at least
where the first term is the probability that , the second term is the probability that both provers choose the same zoom-in dimension, the third term is the probability that , , , the fourth term is the probability that is consistent, the fifth term is the probability that the second prover chooses the a function that extends the first prover’s answer, and the final term is the probability that both provers extend their functions in the same manner.
This proves Lemma 6.1.
7 Proofs of the Main Theorems
7.1 Proof of Theorem 1.3
Theorem 1.3 follows by applying out PCP construction from Section 4 starting with an instance of -Lin from Theorem 2.1. We may take , fix to be a small constant and then take sufficiently large compared to , then and according to Equation (4), and finally take the completeness of the -Lin instance, , so that . It follows that if the original -Lin instance is at least satisfiable, then . On the other hand, if the if the original instance is at most satisfiable for some constant , then by Claim 3.2, the value of the outer PCP is at most
since we take sufficiently large compared to . By Lemma 6.1, it follows that if the original instance is at most satisfiable, then . The proof is concluded as the alphabet size of is .
7.2 Proof of Theorem 1.4
To show quasi-NP-hardness for approximate Quadratic Programming, we rely on the following result due to [ABH+05], who show a reduction from -Prover--Round Games to Quadratic Programming.
Theorem 7.1.
There is a reduction from a 2-Prover-1-Round Games, with graph and alphabets to a Quadratic Programming instance such that:
-
•
The running time of the reduction and the number of variables in is polynomial in and .
-
•
If , then .
-
•
If , then .
We are now ready to prove Theorem 1.4.
Proof of Theorem 1.4.
Starting with a instance of size , which may be arbitrarily large, we take the instance of from Theorem 2.2 of size and field size which is a power of as the starting point of our reduction. Take to be a small constant, and to be a sufficiently large constant relative to in our composed PCP. Finally, we pick and by Equation (4). This yields a -time reduction from to a 2-Prover-1-Game on , with alphabets and the following properties:
-
•
.
-
•
.
-
•
The completeness is at least , where .
-
•
The soundness is at most .
Indeed, the first properties are clear. For the soundness, as the original -Lin instance is at most satisfiable for , we get from Claim 3.2 that
as is sufficiently large relative to , so the soundness of the composed PCP follows by Lemma 6.1. Applying the reduction of Theorem 7.1, we get a reduction to a Quadratic Programming instance such that,
-
•
The running time of the reduction and number of variables in are both
-
•
If the original SAT instance is satisfiable, then
-
•
If the original SAT instance is not satisfiable, then
Note that
whereas the gap between the satisfiable and unsatisfiable cases is . Altogether, this shows that for all there is such that unless , there is no -approximation algorithm for Quadratic Programming on variables. ∎
7.3 Proof of Theorem 1.5
In this section we prove Theorem 1.5, and for that we must first establish a version of Theorem 1.3 for bi-regular graphs of bounded degree. The proof of this requites minor modifications of our construction, as well as the right degree reduction technique of Moshkovitz and Raz [MR10].
7.3.1 Obtaining a Hard Instance of Bipartite Biregular -CSP
We first show that the -Prover--Round game from Theorem 1.3 can be transformed into a hard instance of biregular, bipartite -CSP with bounded degrees. This version may be useful for future applications, and is formally stated below. Call a bipartite -CSP -regular if the left degrees of its underlying graph are all , and the right degrees of its underlying graph are all .
Theorem 7.1.
For every , and sufficiently large , there exist such that given a bipartite -regular -CSP, , with alphabet size , it is -hard to distinguish the following two cases:
-
•
Completeness: ,
-
•
Soundness: .
To prove Theorem 7.1, we start with an instance from Theorem 1.3, and first argue that can be made left regular while almost preserving soundness and completeness by deleting a small fraction of left vertices. We then use the right degree reduction technique of Moshkovitz and Raz [MR10], to obtain a bounded degree bi-regular bipartite -CSP.
Fix , and let to be the -Prover--Round game constructed for Theorem 1.3. Recall that this requires us to choose some large enough relative to , some large enough relative to , and set . We also set , arbitrarily small relative to , and . Finally, we construct our -Prover--Round game from a hard instance of Gap3Lin with the appropriate completeness and soundness, so that it is NP-hard to distinguish between,
It is clear that our -Prover--Round game can equivalently be viewed as an instance of bipartite 2-CSP, so let us analyze the underlying graph. Let denote the set of possible questions to the first prover. Recall that the set of left vertices is,
while the set of right vertices is
The edges and constraints of this graph are generated by a randomized process. Equivalently, there is a weight function over edges . Recall that the weighting is defined by first choosing a uniform , and then according to the process descrbied in Section 4.1.4. For a fixed , define
This is the probability of choosing the conditioned on first choosing . Since we choose uniformly, it follows that
Define the neighborhood of a vertex as,
We will now attempt to remove some left vertices and obtain a bipartite, left-regular -CSP. To this end, we call trivial if there is an equation such that for every basis of of , the points restricted to the variables in are of the form for some .
Claim 7.2.
The fraction of that are trivial is at most .
Proof.
Fix a . Note that it suffices to show that at most vertices of the form are trivial, as for each , there are an equal number of vertices .
Write , where the th equation in contains the variables . Call these three coordinates a block, so that each consists of blocks of consecutive coordinates. Let us bound the fraction of such that is trivial. For , let be the number of blocks where are all of the form for some . Then
where the factor of accounts for the probability that either are not linearly dependent, or . Note that the probability that a specific block is trivial is , hence by linearity of expectation we get that
and therefore
Let be the instance obtained from after removing all trivial from , so that the new instance does not contain any trivial vertices. Let denote the set of left vertices in and let denote the weight function over edges in , which is given by choosing uniformly, and then choosing with probability proportional to . It follows that,
Let be a set of edges in and for notational purposes let us write if is trivial and not in . We have
(14) |
The upper bound is clear from Claim 7.2. For the lower bound, we have .
It follows that has completeness at least and soundness at most . We now bound the size of the neighborhoods in .
Claim 7.3.
For each , we have
Proof.
Let and suppose equation contains variables . Since is not trivial, for each , there must be a point such that the values of restricted to the coordinates of variables are not all equal. Without loss of generality, say that it is for each . It follows that in order to have
must contain an equation with the variable for each . Let denote this set of equations for each . By the regularity assumptions on our 3Lin instance, and for . It follows that must contain exactly one equation from each , and that these form all equations of , so there are at most possible for which there can exist , such that . The lemma follows from the observation that . ∎
Performing the same procedure as in [KR03, Lemma 3.4], we can turn into a bipartite, left-regular -CSP instance without losing too much in completeness or soundness. Let be the upper bound on neighborhood sizes in Claim 7.3.
Claim 7.4.
For any , there is a polynomial time algorithm that takes as input and outputs a bipartite -CSP that is left regular with degree , and has
-
•
Completeness .
-
•
Soundness .
Proof.
We define as follows. For each vertex , do the following. Let be the vertices in . For each , add edges from to . Also add edges from to . It is clear that is left regular with degree , and that for each , there are at most edges between and for , and at most edges between and .
For the completeness and soundness, consider a left vertex . Then it is clear by the above that if a labelling satisfies fraction of constraints involving in , then in the same labelling satisfies at least of the edges incident to . Similarly, if a labelling satisfies at most fraction of constraints involving in , then it satisfies at most edges involving . ∎
Applying Claim 7.4 with , we obtain a bipartite -CSP, , that is left regular with degree , that still has nearly the same completeness and soundness as our original instance . We will now create a -regular bipartite CSP from , by using the degree reduction technique of Moshkovitz and Raz [MR10].
Lemma 7.5.
[MR10] For any parameter , there is a polynomial time algorithm that takes a bipartite, left regular -CSP with left degree, , completeness and soundness , and outputs a bipartite, -regular -CSP completeness and soundness .
We are now ready to complete the proof of Theorem 7.1.
Proof of Theorem 7.1.
7.3.2 Sparsification
In [LM23], Lee and Manurangsi show how to conclude Theorem 1.5 from Theorem 7.1 via a sparsification procedure. We summarize the steps here. Fix the for Theorem 1.5. Set in Theorem 7.1 and let be the resulting hard bipartite -regular 2-CSP, and it is NP-hard to distinguish between.
Now, observe that the degrees of can be multiplied by arbitrary constants by copying vertices.
Lemma 7.6.
[LM23, Lemma 10] For any integers , there is a polynomial time reduction from a bipartite -biregular CSP, , to a bipartite -biregular CSP , such that , and such that the left and right alphabet sizes are preserved.
It is then shown in [LM23] that one can perform a subsampling procedure to , that significantly lowers the degree, while not increasing the soundness or alphabet size too much.
Theorem 7.7.
[LM23, Theorem 11] For any , such that any positive integer , and any sufficiently large positive integers , and , the following holds: there is a randomized polynomial-time reduction from a bipartite -biregular -CSP, , with alphabet size at most , -bounded degree -CSP, , such that, with probability at least ,
-
•
Completeness: ,
-
•
Soundness: If , then
Putting everything together, we can prove Theorem 1.5.
Proof of Theorem 1.5.
Recall the values and from Theorem 1.5. Start with an instance of -CSP from Theorem 7.1 with and sufficiently large alphabet size . Then is -biregular, with sufficiently large alphabet size relative to and . For such a , it is NP-hard to distinguish whether , or .
Applying Lemma 7.6 with yields, in polynomial time, a -biregular -CSP, , with alphabet size and satisfying . Next, applying Theorem 7.7, with to obtain, in randomized polynomial time, a -CSP, , with degree at most such that:
-
•
If , then .
-
•
If , then .
Finally note that,
Thus, by Theorem 7.1 and the randomized polynomial time reduction above, it follows that unless , there is no polynomial time approximation algorithm for 2-CSP with degree at most . ∎
7.4 Proof of Theorem 1.6
Combining our -Prover--Round Game in Theorem 1.3 with [Lae14, Lemma 4] and [Man19, Theorem 1], we obtain improved hardness of approximation results for Rooted -connectivity on undirected graphs, the vertex-connectivity survivable network design problem, and the vertex-connectivity -route cut problem on undirected graphs. The reduction is exactly the same as the reduction therein and we therefore omit the details.
8 Bounding the Number of Successful Zoom-outs of a Fixed Codimension
The goal of this section is to prove Lemma 5.19. Let be the space that we are working in and suppose is a table assigning linear functions to . We assume . Let us review the set up of Lemma 5.19. Recall that we set
Let be a set of codimension -subspaces in of size
For each , let be a linear function such that for at least -fraction of the -subspaces , where , and .
8.1 Step 1: Reducing to a Generic Set of Subspaces
Applying Lemma 5.8, with parameter as defined, we get that there exists a subspace and a set of
subspaces , such that
-
•
Each is contained in and has co-dimension with respect to , where .
-
•
is -generic with respect to .
We remark that this subspace will ultimately be the one used for Lemma 5.19. The remainder of the proof is devoted to finding the linear function , and the set , which will be a subset of above.
8.2 Step 2: Finding Local Agreement
For a subspace and linear assignment to , , let
Likewise, define
The first step of our proof is to find sets and that have strong agreement between them, in the sense of the following lemma. The approach of this first step is similar to that of [IKW12, BDN17, MZ23]. Fix to be a small constant, say .
Lemma 8.1.
There exists a -dimensional subspace , a linear assignment, , to , and sets and such that the following hold:
-
•
.
-
•
.
-
•
Choosing uniformly, and uniformly such that , we have
Proof.
Deferred to Appendix E ∎
As an immediate corollary, we have the following statement. The difference between Corollary 8.2 and Lemma 8.1 is that the former we require the third condition to hold every every , instead of a random , and we also require every to be contained in roughly the same number of .
Corollary 8.2.
Taking and from Lemma 8.1, there is a subset such that the following hold.
-
•
-
•
.
-
•
For every , choosing uniformly such that , we have
-
•
For every ,
where .
Proof.
Take and as guaranteed by Lemma 8.1, so that . We will keep the same , but we remove some ’s from to make the third and fourth items hold.
By Markov’s inequality, at most -fraction of violate the third item. By Lemma 5.16 applied to with parameters , , , , and , we have that
where . It follows that after removing the that do not satisfy the third or fourth condition, we arrive at the desired , which still has measure at least
We fix as well as and as in Corollary 8.2 throughout the rest of the argument.
8.3 Step 3: A Global Set with Local Agreement
The next step is to further refine the set so that the remaining subspaces “evenly cover” a subspace of with codimension . To do this, we will reduce to the case where is global within some zoom-in and zoom-out such that . This is done via the following argument, which we outline below:
-
1.
While is not global, there must be some zoom-in or zoom-out on which it is dense, so consider the restriction to this zoom-in or zoom-out.
-
2.
This increases the measure of , and we may repeat until we have a global set (within some zoom-in, zoom-out combination).
-
3.
By choosing the globalness parameter suitably, we are able to perform the above process in relatively few times until the restriction of that we arrive at is global.
-
4.
As a result, when restricting to the zoom-in, zoom-out combination, the resulting set of subspaces evenly covers a space that is still relatively large in the sense that it contains -fraction of .
For a zoom-in and zoom-out such that , write and , where . Also define
It is clear that each is contained inside of some , so for each we may define the restriction of to as .
Lemma 8.3.
There is a zoom-in and a zoom-out such that the following holds with the notation above. There exists a collection of subspaces of codimension with respect to such that:
-
1.
For some there exists such that .
-
2.
The set is -pseudo-random.
-
3.
Each has codimension inside of and is -generic, with respect to .
-
4.
.
-
5.
For every , choosing uniformly such that , we have
-
6.
For every ,
where
Here the table is assigns linear functions to , and is essentially the original table, i.e
Proof.
Deferred to Appendix E.2. ∎
Finally, as a consequence of pseudo-randomness, we may apply Lemma 2.4, to get that evenly covers .
Lemma 8.4.
Setting , we have that,
Proof.
This is immediate by the pseudo-randomness of and Lemma 2.4. ∎
8.4 Step 4: Local to Global Agreement
Lemma 8.5.
We have
and for every ,
Proof.
Deferred to Section E.3. ∎
Using Lemma 8.5, we conclude the proof of Lemma 5.19 by using ideas from the Raz-Safra analysis of the Plane versus Plane test [RS97]. Define the following graph, , with vertex set and an edge between if and only if . We claim that this graph contains a large clique. To do so, we show that the graph is nearly transitive. For a graph , define
A graph is transitive . It is easy to see that transitive graphs are (edge) disjoint unions of cliques. The following lemma, proved in [RS97], asserts that if is relatively dense and is small, then one could remove only a small fraction of the edges and get a fully transitive graph.
Lemma 8.6.
[RS97, Lemma 2] Any graph can be made transitive by deleting at most edges.
To use lemma 8.6 we first show that the graph we defined is highly transitive.
Claim 8.7.
We have .
Proof.
Fix a that are not adjacent. We claim that they can have at most common neighbor. Suppose for the sake of contradiction that are distinct common neighbors. Then,
and
It follows that and agree on . However, since is -generic, we have
and
so it must be the case that . This contradicts the assumption that and are not adjacent. Thus, any two non-adjacent vertices can have at most common neighbor, and the result follows. ∎
Claim 8.8.
The graph contains a clique of size of size
Proof.
Applying Markov’s inequality and a union bound to Lemma 8.5, we have that with probability at least over and , we have both and . In this case, and agree on at least -fraction of the points in , which is in turn at least -fraction of the points in . As and are linear functions, the Schwartz-Zippel lemma implies that such are adjacent in and that has at least edges.
Let be the clique guaranteed by Claim 8.8 and write . To complete the proof of Lemma 5.19, we will find a linear , such that for all , , and then show that this can be extended to in a manner that is consistent with many of the original ’s for . To this end, first define as follows:
(15) |
Since whenever , it does not matter which is chosen if there are multiple containing . Thus, is well defined and for all .
We next show that is close to a linear function and that this linear function agrees with most of the functions for . To begin, we show that passes the standard linearity test with high probability.
Lemma 8.9.
We have,
Proof.
Note that we have
For every linearly independent, we can let denote the number of containing . The result them follows from Lemma 5.16 with , , , and . We have,
where is an upper bound on the probability that . ∎
Applying the linearity testing result of Blum, Luby, and Rubinfeld [BLR93, Theorem 4.1] we get that is -close to a linear function, say . We will conclude by showing that this is the desired function which agrees with many of the original ’s. To this end, we first show that agrees with many of the ’s that we have (which are restrictions of the original ’s), and then show that can be extended to in a manner that retains agreement with many of the ’s.
Towards the first step, set . We show that choosing randomly, and then a point , it is unlikely that . Define the measure over nonzero points in obtained by choosing uniformly at random and then nonzero uniformly at random. Let be the uniform measure, so . Then is precisely the probability of interest and can be upper bounded using Lemma 5.14, with parameters , codimension ,
(16) |
Lemma 8.10.
We have for at least half of the .
Proof.
By Markov’s inequality and Equation (16) with probability at least , over , we have
and and agree on more than of the points in . Since and are both linear, by the Schwartz-Zippel Lemma that , and the result follows. ∎
We are now ready to finish the proof of Lemma 5.19.
Proof of Lemma 5.19.
Summarizing, we now have linear functions for and a linear function such that . Furthermore, for each , there is a from Lemma 5.19such that , , , and .
Finally, we will extend in a manner so that it agrees with many of these original functions . To this end, recall that satisfies,
and . Therefore, we may choose a random linear function conditioned on and . For any , we have that
Indeed there is a probability that , as we condition on . Then, extending from to , there is at least a probability that is equal to on these extra dimensions. It follows that there is a linear such that for at least
of the pairs from Lemma 5.19. Take these to be the set for Lemma 5.19. As these , they are -generic with respect to and have codimension in . ∎
9 Acknowledgements
We thank the anonymous reviewers for their comments.
References
- [ABH+05] Sanjeev Arora, Eli Berger, Elad Hazan, Guy Kindler, and Muli Safra. On non-approximability for quadratic programs. In 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2005), 23-25 October 2005, Pittsburgh, PA, USA, Proceedings, pages 206–215. IEEE Computer Society, 2005.
- [ALM+92] Sanjeev Arora, Carsten Lund, Rajeev Motwani, Madhu Sudan, and Mario Szegedy. Proof verification and hardness of approximation problems. In 33rd Annual Symposium on Foundations of Computer Science, Pittsburgh, Pennsylvania, USA, 24-27 October 1992, pages 14–23, 1992.
- [AS03] Sanjeev Arora and Madhu Sudan. Improved low-degree testing and its applications. Comb., 23(3):365–426, 2003.
- [BDN17] Amey Bhangale, Irit Dinur, and Inbal Livni Navon. Cube vs. cube low degree test. In Christos H. Papadimitriou, editor, 8th Innovations in Theoretical Computer Science Conference, ITCS 2017, January 9-11, 2017, Berkeley, CA, USA, volume 67 of LIPIcs, pages 40:1–40:31. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2017.
- [Ber00] Piotr Berman. A d/2 approximation for maximum weight independent set in d-claw free graphs. Nordic Journal of Computing, 7(3):178–184, 2000.
- [BKS19] Boaz Barak, Pravesh K. Kothari, and David Steurer. Small-set expansion in shortcode graph and the 2-to-2 conjecture. In Avrim Blum, editor, 10th Innovations in Theoretical Computer Science Conference, ITCS 2019, January 10-12, 2019, San Diego, California, USA, volume 124 of LIPIcs, pages 9:1–9:12. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2019.
- [BL15] Abhishek Bhowmick and Shachar Lovett. The list decoding radius of reed-muller codes over small fields. In Rocco A. Servedio and Ronitt Rubinfeld, editors, Proceedings of the Forty-Seventh Annual ACM on Symposium on Theory of Computing, STOC 2015, Portland, OR, USA, June 14-17, 2015, pages 277–285. ACM, 2015.
- [BLR93] Manuel Blum, Michael Luby, and Ronitt Rubinfeld. Self-testing/correcting with applications to numerical problems. J. Comput. Syst. Sci., 47(3):549–595, 1993.
- [CCK08] Tanmoy Chakraborty, Julia Chuzhoy, and Sanjeev Khanna. Network design for vertex connectivity. In Cynthia Dwork, editor, Proceedings of the 40th Annual ACM Symposium on Theory of Computing, Victoria, British Columbia, Canada, May 17-20, 2008, pages 167–176. ACM, 2008.
- [CDGT22] Julia Chuzhoy, Mina Dalirrooyfard, Vadim Grinberg, and Zihan Tan. A new conjecture on hardness of low-degree 2-csp’s with implications to hardness of densest k-subgraph and other problems. CoRR, abs/2211.05906, 2022.
- [CGM13] Marek Cygan, Fabrizio Grandoni, and Monaldo Mastrolilli. How to sell hyperedges: The hypermatching assignment problem. In Sanjeev Khanna, editor, Proceedings of the Twenty-Fourth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2013, New Orleans, Louisiana, USA, January 6-8, 2013, pages 342–351. SIAM, 2013.
- [Cha16] Siu On Chan. Approximation resistance from pairwise-independent subgroups. Journal of the ACM (JACM), 63(3):1–32, 2016.
- [CW04] Moses Charikar and Anthony Wirth. Maximizing quadratic programs: Extending grothendieck’s inequality. In 45th Symposium on Foundations of Computer Science (FOCS 2004), 17-19 October 2004, Rome, Italy, Proceedings, pages 54–60. IEEE Computer Society, 2004.
- [DFRR23] Pavel Dvořák, Andreas Emil Feldmann, Ashutosh Rai, and Paweł Rzǎżewski. Parameterized inapproximability of independent set in h-free graphs. Algorithmica, 85(4):902–928, 2023.
- [DKK+18] Irit Dinur, Subhash Khot, Guy Kindler, Dor Minzer, and Muli Safra. Towards a proof of the 2-to-1 games conjecture? In Ilias Diakonikolas, David Kempe, and Monika Henzinger, editors, Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2018, Los Angeles, CA, USA, June 25-29, 2018, pages 376–389. ACM, 2018.
- [DKK+21] Irit Dinur, Subhash Khot, Guy Kindler, Dor Minzer, and Muli Safra. On non-optimally expanding sets in grassmann graphs. Israel Journal of Mathematics, 243(1):377–420, 2021.
- [EKL23] David Ellis, Guy Kindler, and Noam Lifshitz. An analogue of bonami’s lemma for functions on spaces of linear maps, and 2-2 games. In Barna Saha and Rocco A. Servedio, editors, Proceedings of the 55th Annual ACM Symposium on Theory of Computing, STOC 2023, Orlando, FL, USA, June 20-23, 2023, pages 656–660. ACM, 2023.
- [EKL24] Shai Evra, Guy Kindler, and Noam Lifshitz. Polynomial bogolyubov for special linear groups via tensor rank, 2024. arXiv:2404.00641.
- [Fei98] Uriel Feige. A threshold of ln n for approximating set cover. J. ACM, 45(4):634–652, 1998.
- [FGL+91] Uriel Feige, Shafi Goldwasser, László Lovász, Shmuel Safra, and Mario Szegedy. Approximating clique is almost np-complete (preliminary version). In 32nd Annual Symposium on Foundations of Computer Science, San Juan, Puerto Rico, 1-4 October 1991, pages 2–12, 1991.
- [Gop13] Parikshit Gopalan. A fourier-analytic approach to reed-muller decoding. IEEE Trans. Inf. Theory, 59(11):7747–7760, 2013.
- [GRS00] Oded Goldreich, Ronitt Rubinfeld, and Madhu Sudan. Learning polynomials with queries: The highly noisy case. SIAM J. Discret. Math., 13(4):535–570, 2000.
- [Hås96] Johan Håstad. Clique is hard to approximate within n. In 37th Annual Symposium on Foundations of Computer Science, FOCS ’96, Burlington, Vermont, USA, 14-16 October, 1996, pages 627–636. IEEE Computer Society, 1996.
- [Hås01] Johan Håstad. Some optimal inapproximability results. J. ACM, 48(4):798–859, 2001.
- [IKW12] Russell Impagliazzo, Valentine Kabanets, and Avi Wigderson. New direct-product testers and 2-query PCPs. SIAM J. Comput., 41(6):1722–1768, 2012.
- [KMS17] Subhash Khot, Dor Minzer, and Muli Safra. On independent sets, 2-to-2 games, and grassmann graphs. In Hamed Hatami, Pierre McKenzie, and Valerie King, editors, Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2017, Montreal, QC, Canada, June 19-23, 2017, pages 576–589. ACM, 2017.
- [KMS23] Subhash Khot, Dor Minzer, and Muli Safra. Pseudorandom sets in grassmann graph have near-perfect expansion. Annals of Mathematics, 198(1):1–92, 2023.
- [KP06] Subhash Khot and Ashok Kumar Ponnuswami. Better inapproximability results for maxclique, chromatic number and min-3lin-deletion. In Michele Bugliesi, Bart Preneel, Vladimiro Sassone, and Ingo Wegener, editors, Automata, Languages and Programming, 33rd International Colloquium, ICALP 2006, Venice, Italy, July 10-14, 2006, Proceedings, Part I, volume 4051 of Lecture Notes in Computer Science, pages 226–237. Springer, 2006.
- [KR03] Subhash Khot and Oded Regev. Vertex cover might be hard to approximate to within 2-\varepsilon. In 18th Annual IEEE Conference on Computational Complexity (Complexity 2003), 7-10 July 2003, Aarhus, Denmark, page 379. IEEE Computer Society, 2003.
- [KS13] Subhash Khot and Muli Safra. A two-prover one-round game with strong soundness. Theory Comput., 9:863–887, 2013.
- [Lae14] Bundit Laekhanukit. Parameters of two-prover-one-round game and the hardness of connectivity problems. In Chandra Chekuri, editor, Proceedings of the Twenty-Fifth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2014, Portland, Oregon, USA, January 5-7, 2014, pages 1626–1643. SIAM, 2014.
- [LM23] Euiwoong Lee and Pasin Manurangsi. Hardness of approximating bounded-degree max 2-csp and independent set on k-claw-free graphs. CoRR, abs/2309.04099, 2023.
- [Man19] Pasin Manurangsi. A note on degree vs gap of min-rep label cover and improved inapproximability for connectivity problems. Inf. Process. Lett., 145:24–29, 2019.
- [Meg01] Alexandre Megretski. Relaxations of quadratic programs in operator theory and system analysis. In Systems, Approximation, Singular Integral Operators, and Related Topics: International Workshop on Operator Theory and Applications, IWOTA 2000, pages 365–392. Springer, 2001.
- [MR10] Dana Moshkovitz and Ran Raz. Two-query PCP with subconstant error. J. ACM, 57(5):29:1–29:29, 2010.
- [MZ23] Dor Minzer and Kai Zheng. Approaching the soundness barrier: A near optimal analysis of the cube versus cube test. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 2761–2776. SIAM, 2023.
- [NRT99] Arkadi Nemirovski, Cornelis Roos, and Tamás Terlaky. On maximization of quadratic form over intersection of ellipsoids with common center. Mathematical programming, 86(3):463–473, 1999.
- [Nut12] Zeev Nutov. Approximating minimum-cost connectivity problems via uncrossable bifamilies. ACM Transactions on Algorithms (TALG), 9(1):1–16, 2012.
- [Rao11] Anup Rao. Parallel repetition in projection games and a concentration bound. SIAM J. Comput., 40(6):1871–1891, 2011.
- [Raz98] Ran Raz. A parallel repetition theorem. SIAM J. Comput., 27(3):763–803, 1998.
- [RS97] Ran Raz and Shmuel Safra. A sub-constant error-probability low-degree test, and a subconstant error-probability pcp characterization of np. In Proceedings of the twenty-ninth annual ACM Symposium on Theory of Computing, pages 475–484, 1997.
- [Tre14] Luca Trevisan. Inapproximability of combinatorial optimization problems. Paradigms of Combinatorial Optimization: Problems and New Approaches, pages 381–434, 2014.
- [TW23] Theophile Thiery and Justin Ward. An improved approximation for maximum weighted k-set packing. In Nikhil Bansal and Viswanath Nagarajan, editors, Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms, SODA 2023, Florence, Italy, January 22-25, 2023, pages 1138–1162. SIAM, 2023.
Appendix A Proofs of Lemmas 2.3 and 2.4
In this section we prove Lemmas 2.3 and 8.4. The proofs of these lemmas requires tools from [EKL23, EKL24] regarding Fourier analysis over the Bilinear Scheme.
A.1 Fourier Analysis over the Bilinear Scheme
The key to proving Lemma 2.3 is a level-d inequality for indicator functions on the Bilinear Scheme due to Evra, Kindler, and Lifshitz [EKL24]. In order to use the result of [EKL24], however, we first give some necessary background for Fourier analysis over the Bilinear Scheme, and describe the analogues of zoom-ins, zoom-outs, and pseudorandomness. The latter is done in [EKL23, EKL24]. After doing so, we must then find a suitable map from the Grassmann graph to the Bilinear Scheme that (1) preserves the edges of our original bipartite inclusion graph between -dimensional and subspaces, and (2) maps zoom-ins and zoom-outs in the Grassmann graph to their analogues over the Bilinear Scheme.
The Bilinear Scheme:
Let be the set of matrices over . One can define a graph over that is similar to the Grassmann graphs by calling adjacent if for some . A graph of this form are often referred to as the Bilinear Scheme. For our purposes, we will need to work with a bipartite version of this graph between and . We equip the space with the following inner product:
where the distribution taken over is uniform. Let be a primitive th root of unity, where recall is the characteristic of . For and , let where is the trace map. Then, the characters, over all , given by
form an orthonormal basis of . As a result, any can be expressed as,
where . The level component of is given by
If a function only consists of components up to level , i.e. for all , then we say is of degree .
We now describe the analogues of zoom-ins and zoom-outs on . We also define the analogous notion of -pseudo-randomness for Boolean functions over , and we begin by defining the analog of zoom-ins.
Definition A.1.
A zoom-in of dimension over is given by -pairs of vectors where each and each . Let and denote the matrices whose th columns are and respectively. Then the zoom-in on is the set of such that , or equivalently, for .
Next, we define the analog of zoom-outs.
Definition A.2.
A zoom-out of dimension is defined similarly, except by multiplication on the left. Given and , whose rows are given by and respectively, the zoom-out is the such that , or equivalently, for .
Let denote the intersections of the zoom-in on and the zoom-out on . The codimension of is the sum of the number of columns of and the number of rows of , which we will denote by and . For a zoom-in and zoom-out pair and a Boolean function , we define to be the restriction of which is given as
When , we say that the restriction is of size . We define -pseudo-randomness in terms of the -norms of restrictions of of size . Here and throughout, when we consider restricted functions, the underlying measure is the uniform measure over the corresponding zoom-in and zoom-out set .
Definition A.3.
We say that an indicator function is -pseudorandom if for all zoom-in zoom-out combinations such that , we have
We note that for Boolean functions , , and hence the definition above generalizes the definition we have for Boolean functions.
Definition A.4.
We say that an indicator function is -pseudo-random if for all such that , we have,
The following result is a combination of two results form [EKL24]. Roughly speaking, it states that if a Boolean function is pseudo-random, then its degree parts are pseudo-random for .
Lemma A.5.
[EKL24, Theorem 5.5 + Proposition 3.6] Let be a power of and let be a function that is -pseudo-random. Then is -pseudo-random for all .
Proof.
This lemma does not actually appear in [EKL24], but it is easy to derive by combining Theorem 5.5 with Proposition 3.6 therein. In [EKL23, EKL24], the authors introduce an additional notion of generalized influences and having small generalized influences. We refrain from defining these notions explicitly as it is slightly cumbersome, but roughly speaking, one defines a Laplacian for each zoom-in, zoom-out combination, so that having small generalized influences means that upon applying these Laplacians on , the -norm squared of the resulting function never exceeds .
Lastly, we need the following global hypercontrativity result also due to [EKL24].999We remark that earlier results [EKL23] showed similar statement for -norms, i.e. the case that , and the result below follows by a form of induction on . That is, one starts with and concludes via applying the case that the function and it is pseudo-random. Then one apply the case on to conclude that is pseudo-random and so on.
Theorem A.6.
[EKL24, Theorem 1.13] Let be a power of and let be a function of degree that is -pseudo-random. Then,
Combining Lemma A.5 and Theorem A.6, we arrive at the following result which bounds the -norm of the level component of pseudo-random indicator functions. This result will be the key to showing an analogue of Lemma 2.3 over the Bilinear Scheme.
Theorem A.7.
Let be a power of . Then if is -pseudo-random, we have
for all .
Proof.
Suppose is -pseudo-random, let be a power of , and fix a . Since , we also have that is -peudorandom. Therefore for any size restriction of , ,we have,
Thus, is -pseudo-random, and by Lemma A.5 it follows that is -pseudo-random. Clearly, is degree , so applying Theorem A.6 we get that,
where we also use the fact that because is an -pseudorandom Boolean function. Taking the -th root of the above inequality completes the proof. ∎
A.2 An Analog of Lemmas 2.3 for the Bilinear Scheme
With Theorem A.7 in hand we can show an analogue of Lemma 2.3 for basis invariant functions over the Bilinear Scheme. To do so, we first define what we mean by basis invariant functions, then present an analogue of the adjacency operator (which is originally defined for functions over subspaces) over the Bilinear Scheme, which we denote by , and finally show that the previously described characters are eigenoperators of , where is an operator that acts as the adjoint of on basis invariant functions.
For a function , say that is basis invariant if for any full rank . We first show that the level component of a basis invariant function is also basis invariant. The following identity regarding the characters will be useful.
Lemma A.8.
For any , any , and any matrix we have,
Proof.
Letting denote the columns of and denote the entries of , we have,
Lemma A.9.
Let and let be basis invariant. Then for any that is full rank, we have .
Proof.
For any matrix full rank we have
where we use that is basis invariant in the third transition and Lemma A.8 in the fourth transition. Setting gives the result. ∎
Using Lemma A.9, we can show that the level component of a basis invariant function is also basis invariant.
Lemma A.10.
If is basis invariant, then is basis invariant as well for any .
Proof.
We now define the following two operators which will be the analogues of and over the bilinear scheme. The first is , given by:
In words, the operator averages over extensions of the matrix to an matrix by adding to it random columns. The next is given by:
Strictly speaking, is not the adjoint of ; however, for the case where is basis invariant, acts as the adjoint of in the following sense.
Lemma A.11.
For that is basis invariant and , we have
Proof.
Let be the matrix whose restriction to the first rows is the identity matrix and whose remaining rows are all . We have
where in the second transition we used the fact that is basis invariant, and in the third one we made a change of variables . Now note that is the matrix restricted to its first columns and hence in the final distribution, is a uniformly random matrix in with rank . It follows that,
We will want to understand the operator , and towards this end we define the operator
The reason for introducing is that, as the following lemma shows, it acts the same as on basis invariant functions, but is easier to work with. This is due to the reason it is an averaging operator with respect to some Cayley graph over , and therefore each character is an eigenvector of and the eigenvalues have an explicit formula. These facts are shown in the next two lemmas respectively.
Lemma A.12.
If is basis invariant, then .
Proof.
By definitions
We can also view as being sampled as follows. Choose with linearly independent columns, extend to a matrix with linearly independent columns randomly by adding columns on the right, sample a random matrix , and output,
Furthermore, under this distribution, it is clear that is a uniformly random matrix with linearly independent columns. Therefore,
where we are using the fact that is basis invariant and is invertible. In the last expectation, note that the distribution over is the same as that over where is uniformly random, and is uniformly random conditioned on having linearly independent rows. More precisely, it is equal to where , and is the last rows of . It follows that
The following lemma gives upper bound on the eigenvalues of .
Lemma A.13.
Suppose that . If , then is an eigenvector of of eigenvalue . Otherwise, if , is an eigenvector of of eigenvalue which is at most in absolute value.
Proof.
Fix . We argued earlier that is an eigenvector of , and we denote the corresponding eigevalue by . If the statement is clear, so we assume that henceforth.
Find of full rank so that where are linearly independent. Thus, as the distribution of is invariant under multiplying by from the right, we get that
and we may assume that for linearly independent to begin with. Applying symmetry again, we conclude that
and interchanging the order of expectations we get that
Denote , and inspect these vectors.
Claim A.14.
If for some , then
Proof.
We first claim that if are chosen uniformly, then the left hand side is , or equivalently
To see this, it suffices to show that takes every value in with equal probability, and we focus on showing this. Fix such that and suppose the th entry, is nonzero. We can fix all entries of the uniformly except for , and then for each , there is exactly one choice of that will result in .
Thus, if we took the distribution over to be uniformly and independently chosen, then the magnitude of the above expectation would be . Hence, we conclude that the above expectation is at most twice the probability randomly chosen are not linearly independent, which is at most . ∎
By Claim A.14 we conclude that , and we next bound this probability. Recalling the definition of , we have that
Consider the minor of and call it . First we upper bound the probability that . Note that the distribution of is the same as of where is a random invertible matrix. Thus, has the same distribution as of , and the probability that is at most
It remains to bound the probability that are all in the case that . In this case, assume without loss of generality that the first column of is non-zero. Thus, it follows that over the randomness of , the vector is uniformly chosen from , and so the probability it is the all vector is at most . Combining, we get that . ∎
Finally, using Lemma A.8 again, we can show that does not increase the level of a function,
(17) |
and obtain a useful identity for decomposing inner products.
Lemma A.15.
Let and . Then,
As a consequence we also have
Proof.
We are now ready to state and prove an analog of Lemma 2.3 for basis invariant functions on the Bilinear scheme.
Lemma A.16.
Let and be basis invariant indicator functions with . If is pseudo-random and basis invariant, then for all that are powers of , we have
Proof.
Using the degree decomposition of and Lemma A.15, we can write
We first bound the contribution from terms in the summation with using Cauchy-Schwarz. For ,
where the third transition uses Lemma A.12 and the fact that is basis invariant by Lemma A.10, and finally the fourth transition uses Lemma A.13. Thus, the total contribution from the terms is
A.3 Reduction to the Bilinear Scheme
We are now ready to prove Lemma 2.3. As in the statement of Lemma 2.3, let and be Boolean functions, and suppose that is -pseudo-random. Define the Boolean functions , by
and
We note that and are basis invariant functions. Next, we will prove that is pseudo-random, and towards this end we begin with the following lemma that simplifies the type of zoom-ins and zoom-out combinations we have to consider for .
Lemma A.17.
For any such that and is nonempty, there are linearly independent columns of , say and a subset of linearly independent rows of , say , such that , and
where is the matrix whose first columns are , and remaining columns are .
Proof.
Let and . First note that we can assume that the columns of and respectively are both nonzero and linearly independent. Indeed, otherwise say , then either , in which case the th columns of and can be removed, or and is an empty set. Otherwise, if, say, , then either and again we can ignored the th columns, or and consists of matrices whose columns are not linearly independent. In this case is identically and the statement is trivially true. Similarly, if the columns of are not linearly independent, then consists of matrices whose columns are not linearly independent, and again is identically . Finally, if the columns of are not linearly independent, then either is empty or there must be some such that both and are linear combinations of the other columns in and respectively, with the same coefficients. In this case, we can remove the th columns of and without changing .
Now suppose that the columns of and are nonzero and linearly independent, and let be a full rank matrix such that for . Let denote the last columns of , and let denote the first columns of . Since we assumed is nonempty, we must have . Then by the fact that is basis invariant we get that
To complete the proof, we show to reduce to the case that is the zero matrix. First note that, using the same reasoning as we did for and , we can assume that the nonzero rows are linearly independent and the rows of are linearly independent
Suppose that are the nonzero (and linearly independent) rows of , while the remaining rows are . Let be the first a rows of , which are nonzero, let denote the first rows of , and let denote rows through of . For any with linearly independent rows let be the full rank matrix such that . Then, for any linearly independent ,
In the second transition, we used the fact that is basis invariant and multiplied its input by the matrix whose top left minor is the identity, its bottom right is , and the rest of the entries are . Since the above holds for any with -linearly independent rows, letting denote the event that has linearly independent rows, it follows that
and
Finally since (as it is the probability of choosing linearly independent vectors in ), we have,
and the proof is concluded. ∎
As an immediate consequence of Lemma A.17
Lemma A.18.
If is -pseudo-random then is -pseudo-random.
Proof.
Fix any and such that . Using Lemma A.17, there are linearly independent and with linearly independent rows such that
where in the second transition we are using the fact that for all such that , and in the third transition we are using the definition of . We will bound the final term by using the pseudo-randomness of .
Choosing uniformly conditioned on , and , we claim that is a uniformly random -dimensional subspace in , where is the codimension subspace that is dual to the rows of . To see why, first note that it is clear . Additionally, each has an equal number of such that
and therefore has an equal chance of being selected. It follows that choosing uniformly conditioned on , and , is a uniformly random -dimensional subspace in . As a result,
where in the last transition we use the fact that is -pseudo-random and . ∎
Next, we note that the values of and are similar.
Lemma A.19.
We have
Proof.
We have,
To finish the proof, notice that in the conditional distribution in the last term, is a uniform and is a uniform such that . Therefore,
A.4 Proof of Lemma 2.4
We will show that if a set of -dimensional subspaces is pseudo-random, then it must “evenly cover” the space in the sense that there are very few points such that significantly deviates from . We will require the following result from [EKL24].
Theorem A.20.
[EKL24, Theorem 1.12] If is a Boolean function which is -global, then for all powers of that are powers of it holds that
Lemma A.21.
Let have for some large constant , and set . If is -pseudo-random for some , then
Proof.
Let , let be the indicator function for , and let be the associated function given by
By Lemma A.18, is -pseudo-random. For any point , we have
so it follows that
Thus
(19) |
We will now relate this quantity to the level one weight of and apply Lemma A.7 to bound the level one weight of . Note that
Now note that if is not the zero vector, and if is the zero vector. Thus,
and using the fact that ,
Therefore, we get by (19) that
Next since is basis invariant, using Lemma A.8, we have that for all that are not all zero,
It follows that
Using Theorem A.20 with , along the fact that is -pseudo-random, we get
for any that is a power of . By Markov’s inequality it follows that
where we take to be a power of large enough so that . Dividing by finishes the proof. ∎
Appendix B Proof of Theorem 5.2
In order to prove Theorem 5.2 we will find the subspaces, , one at at time by using Theorem 5.1. We let denote the set of all ’s collected thus far. Each time a new subspace is added to , we randomize the assignment for all -dimensional . At a high level, the effect of this randomization is that there is only a little agreement between any linear function and the assignments on subspaces containing , thus these entries are essentially “deleted”.
More formally, we construct the set of subspaces as follows. Initially set , , and . Recall that initially and are -consistent for . While and are at least -consistent, do the following.
-
1.
Let be subspaces guaranteed by Theorem 5.1. That is, and satisfy and there exists linear such that
-
2.
Set .
-
3.
Set .
-
4.
For each independently, choose uniformly among all linear functions on .
We have the following claim regarding the re-assignment phase.
Claim B.1.
With probability over the random assignment on , for every such that and at least one of the following holds:
-
1.
Less than -fraction of are in ,
-
2.
For every linear function ,
Proof.
Note that the first item has nothing to do with the random assignment over , so we need only show that if the first item is false then the second item must be true. Suppose that the first item does not hold and that contains at least -fraction of .
Fix with the parameters above and let , and suppose that the first item does not hold. We will show that in this case, the second item holds.
For each , let denote the indicator variable that takes value if and otherwise. The expectation of is , and by a Chernoff bound, the desired probability is bounded by
By assumption, . Thus, using a union bound over all , the probability that there exist a bad triple is at most,
We now analyze the process. Note that using Chernoff’s bound, with probability over the randomization step, the probability drops from at least to at most . In that case, the measure of increases by at least
Doing a union bound over the steps, it follows that with probability the process terminates within steps.
Note that it is possible that the same subspace is added multiple times (with different zoom-outs) in the process above, so we clarify that is considered as a set without repeats. Also note that with probability , for each , and found in the process it holds that
(20) |
(the point being is that the agreement now is compared to the original and not to ). Indeed, considering the step and were found, had agreement at least with on at that point, and by Claim B.1 with probability at most of that agreement came from . Thus, by union bound over all of the steps, with probability it follows that (20) holds for every and found throughout the process.
The following claim shows that at the end of the process the number of ’s found in the process is large, thereby finishing the proof of Theorem 5.2.
Claim B.2.
There exists some such that contains at least a -fraction of all -dimensional subspaces.
Proof.
At the end of the process, the consistency has dropped by at least , so the probability over edges that was reassigned must be at least . For each , let be the number of of dimension in .
For each of dimension , the fraction of -dimensional ’s that are reassigned due to being added to is at most the fraction of -dimensional subspaces that contain . This is,
It follows that there must be some such that
Rearranging this inequality, we get that
Thus there exists an such that contains at least a -fraction of all -dimensional subspaces. ∎
Appendix C The Covering Property
Fix a question to the first prover. Recall that we set
where is some small constant close to and set , and recall that the distributions and are defined as follows.
-
•
Choose uniformly.
-
•
Output the list .
-
•
Choose according to the Outer PCP.
-
•
Choose uniformly, and lift these vectors to by inserting ’s into the missing coordinates.
-
•
Choose uniformly, and set for .
-
•
Output the list .
C.1 Proof of Lemma 5.4
For , let us view as the rows of a matrix, and split the columns of this matrix into blocks - each consisting of consecutive columns. Then let be the number of blocks where at least two of the columns are equal, and set . The idea is that should correspond to the number of equations where we drop variables in the Outer PCP, while is the probability that a fixed block has at least two columns equal to each other. Also let be the number of blocks where all columns are equal, and let , be the probability that a fixed block has all three columns equal.
We define the set as follows. Set
(21) |
(22) |
and .
By a Chernoff bound,
where recall that . Also, by our setting of , we have , while , so the same Chernoff bound holds for ,
so . Indeed, the actual expectation of under is and this differs from by only .
For the measure of we have,
where in the middle term, the first factor is the number of ways to choose blocks and the second factor is the probability that all of these blocks have all three columns equal. Similarly,
Putting everything together, we get that
(23) |
We next show that the probability measures and assign roughly the same measure to each . Fix . It is clear that , where we use . Let and . Then,
(24) |
In the first equality, the first term is the probability of choosing the blocks that have three all distinct columns. Then, is the probability that no variables are dropped, and is the probability of choosing those three particular ’s in that block. The second term is the probability of choosing the blocks that have exactly two equal columns. Then, is the probability of having no variables dropped and choosing the three ’s, and is the probability of first having the variable dropped in the column that is not equal to the other two, and then choosing the correct values for the remaining two column values.
We will first show
Using Equation (C.1),
where in the fourth transition we use the bound , which holds for all real numbers such that is sufficiently small. For our uses, and are and , and both are sufficiently small. In the last transition we use the fact that .
Now write and let us analyze the first two terms in the last line. Plugging this in and using the definition of we get that
(25) |
Plugging this back into the above,
Plugging in our values for and , the first term on the right hand side above is . Using , the second term on the right hand side is at least . Finally, the last term is at least . Overall, we get that for large enough
For the other direction, we show
in nearly the same fashion. First note that
(26) |
By Equation (C.1), we have
where in the first transition we use Equation (26), and in the fourth transition we use the fact that . Writing and using Equation (25) again and using we have,
Appendix D List Decoding Bound
Lemma 5.20 follows directly from a generic list decoding bound of [GRS00, Theorem 15], which we state below for convenience.
Theorem D.1.
[[GRS00, Theorem 15]] Let be a code with alphabet size , blocklength , and relative distance . Let and , where . Suppose that are distinct codewords from that satisfy for all . If,
then
Proof of Lemma 5.20.
Let , let , and define a code with alphabet by:
Note that for distinct we have that for at most -fraction of . Thus, the relative distance of is . We would like the table corresponds to a word, say, , and correspond to codewords in , say . A slight issue is that is only defined over -dimensional subspaces of , while has an entry for every -tuple of points in . To resolve this, note that nearly every -tuple of points combined with span an . Thus, define as follows. If are linearly independent, then let be the span of and define
Otherwise, define arbitrarily. Note that the fraction of tuples such that are not linearly independent is at most,
so nearly all of the entries in correspond to table entries in . For the functions we define corresponding to by
As each agrees with on at least -fraction of the entries, we have that and agree on at least -fraction of the entries such that are linearly independent, so
for each . Finally, note that the alphabet size of is . To bound , we can apply Theorem D.1 with , , and . We first note that the condition of Theorem D.1 is indeed satisfied,
Thus Theorem D.1 implies that . ∎
Appendix E Missing Proofs from Section 8
This section contains the missing proofs from Section 8, and we begin by recalling some notation. We denote by the measure of a collcetion of subspaces , where and will always be clear from context. Furthermore, we use to denote the measure of restricted to the subspaces containing for some subspace low-dimensional subspace , i.e.
Likewise, when is a constant co-dimension subspace, we define
It will always be clear from the size of or in context which of the above definitions we are referring to. We also use
to denote the measure of restricted to the zoom-in of and the zoom-out of . Finally, throughout this section, for some subspace , and a set of constant codimension subspaces , we will let .
E.1 Proof of Lemma 8.1
Recall that is a set of subspaces of codimension inside of that is -generic with respect to . For each -dimensional subspace and linear assignment, , to , define
where in both probabilities and are fixed, and is chosen uniformly and is chosen uniformly. The intention behind these values is that for a fixed , the quantity should reflect how much disagreement there is between the table and the functions for , on subspaces , while should reflect the size of . Note that if and , then by definition we already have . Therefore we would expect that in fact and also agree on - which is only larger than by dimensions - and for most , the value is small. On the other hand, for each , there are at least a -fraction of for which , so we should also expect to be for a non-trivial fraction of . In the following claim, we formalize this intuition and show that there indeed exists an for which is small, is large, and additionally the set is large.
This idea of looking for such was first introduced in [IKW12] where they call these -excellent and was then used again in [BDN17, MZ23] to analyze lower dimensional subspace versus subspace tests, which is similar in spirit to what we are ultimately trying to show in Lemma 5.19.
Claim E.1.
There exists and such that:
-
•
.
-
•
.
-
•
.
Proof.
Consider the following process which outputs such that is uniform in , is uniform, is uniform, and is the assignment of to , i.e .
-
1.
Choose with probability proportional to .
-
2.
Choose uniformly.
-
3.
Choose a -dimensional subspace uniformly conditioned on .
Notice that the distribution of above is equivalent to that of choosing uniformly and uniformly. Moreover, only if , as and must agree on in order to agree on . Therefore,
(27) |
On the other hand,
(28) |
Here the distribution of is proportional to the sizes and the second inequality is by the Schwartz-Zippel lemma. Indeed, by the Schwartz-Zippel lemma, and can agree on at most -fraction of points in . Therefore, the middle term is bounded by the probability that uniformly random, linearly independent points are all chosen in this -fraction.
From a dyadic-partitioning of Equation (27), it follows that there exists a such that
(29) |
By Markov’s inequality on Equation (28)
(30) |
It follows that for at least -fraction of ’s (under the measure induced by step 1 of the sampling procedure above), we have both and .
Next we wish to argue that for most of these ’s, is large. First note that the total number of ’s is at most . For a fixed , the probability that it is chosen is precisely,
Thus, by a union bound,
(31) |
Putting Equations (29), (30), and (31), together, it follows that with probability at least
over , we have, , , and , which establishes the claim. ∎
Taking the given by Claim E.1, it almost looks like Lemma 8.1 is satisfied. However, notice that while the probability of interest for the third item there looks similar to , it has a different distribution over and . Indeed, there, the distribution first chooses , then , whereas for , we are first choosing , and not conditioning being in the set . Intuitively, we expect something like the following to hold,
and be done. These equalities are not actually true however, so the bulk of the transition from Claim E.1 to Lemma 8.1 is in formalizing this chain of equalities and converting from the distribution of to that required by the third item of Lemma 8.1 without losing too much.
Proof of Lemma 8.1.
Fix an such that Claim E.1 holds and define and accordingly. Let . In order to lower bound the measure , we will apply Lemma 5.14 on the collection of subspaces with parameters , . Then the measure over in Lemma 5.14 is precisely that obtained by choosing uniformly, and then uniformly. Thus is precisely the distribution used to define , so by Lemma 5.14
and the first two conditions of Lemma 8.1 are satisfied.
To show the third condition, it will be helpful to have in mind the bipartite graph with parts and and edges if . For each and define the following degree-like quantities:
-
•
,
-
•
,
-
•
,
-
•
.
Also let , where the is arbitrary (the value is the same regardless which we pick). Then and the probability that we are interested in bounding can be expressed as:
Since and , we have
(32) |
and
(33) |
By Lemma 5.16 and the very loose bound , we have
E.2 Proof of Lemma 8.3
Take the and guaranteed by Corollary 8.2, and recall is the ambient space and . As this section is more involved, we restate Lemma 8.3 as well as its setting. Recall that for a zoom-in and zoom-out such that , we write and , where . Now let . It is clear that each is contained inside of some , so for each , we may define .
See 8.3
As a step towards Lemma 8.3, we first show Lemma E.2, which finds the basic items required for Lemma 8.3, modulo a few minor alterations.
Lemma E.2.
We can find a zoom-in and a zoom-out such that , such that the following hold.
-
•
.
-
•
Letting , we have in and is -global in .101010By -pseudo-random in we mean that does not increase its fractional size to when restricted to any zoom-in containing or any zoom-out contained in .
Proof.
Set , , , and . Now do the following.
-
1.
Set , and initialize as above.
-
2.
If is -global inside of , then stop.
-
3.
Otherwise, there exist such that, , , and .
-
4.
Set , , and .
-
5.
Set .
-
6.
Increment by and return to step .
Suppose this process terminates on iteration . We claim that taking , , , for each , and satisfies the requirements of the lemma.
Next notice that by construction . Therefore, we perform at most iterations before stopping, so . By construction is -global in and has fractional size in . Moreover, , so the conditions of the lemma are satisfied. ∎
Take and given by Lemma E.2. Before moving on the the straightforward derivation of Lemma 8.3, define
Proof of Lemma 8.3.
We now construct the and that satisfy Lemma 8.3. Let be a subspace such that , set , and let
For each , we will use to denote the corresponding subspace such that , and the fact that the correspondence,
is a bijection between and . Recall that we abuse notation and let to denote both the original table on -dimensional subspaces, as well as the new table on , given by . It will always be clear, based on the argument in , which we are referring to. We obtain in a similar way as , however, some care will be needed to ensure that it is -generic. First, set , then take to be the maximal subset of that is -generic with codimension with respect to . Finally, set
and set . Summarizing, we have the following chain of relations,
(35) | ||||||||
(36) |
which will be helpful to refer back to. The double arrow transitions are bijections, while in the single arrow transitions subspaces are being removed. The second line shows what a generic member of each set looks like, where are the original subspaces in . We remark that we allow to be a multiset. For each , we define
It is clear from Equation (35) that . We now verify that the six properties of Lemma 8.3 hold.
Property 1. The subspaces of are of dimension inside , and
Also, is the same as the measure of inside due to the bijection between and . Therefore which is at least by the second part of Lemma E.2.
Property 2. Since does not increase its measure to on any zoom-in containing or zoom-out inside , it follows that is -pseudo-random.
Property 3. By construction, is -generic inside of . Since , and all contain , it follows that is -generic inside of .
We verify property 4 using properties 5 and 6, so we save it for last.
Property 6. Fix an with corresponding . We will first show that for all , the value is nearly the same, and in particular is nearly equal to . This has the secondary consequence that is large. By Lemma 5.16 applied to the -generic set of subspaces and we can conclude that Property 5 holds for most (and in particular at least one) subspaces of . We can then conclude that the same holds for all .
Towards showing that all are nearly the same, note that
and additionally
(37) |
so
Since we already have bounds on from the fourth part of Corollary 8.2, it is sufficient to upper bound . Since is -generic, by Lemma 5.13, the set can be made -generic by removing at most of the subspaces in , so
(38) |
Using the fourth part of Corollary 8.2, we get the secondary consequence that
Using Lemma 5.16, we can show that for at least one ,
(39) |
and combining this with Equation (38) shows that Property 6 holds for all .
Property 5. We encourage the reader to refer back to the chain of relations in Equation (35) for this part. At a high level, we will start with a probability regarding at the right end of the chain, and gradually move leftwards and relate this to a probability regarding - which we have a bound on from the fourth item of Corollary 8.2. To start, note that
The first transition is simply due to the fact that . For the second transition we use the fact that there is a one-to-one correspondence between and . For this pair , , we have . Therefore the from the second probability can be sampled by first choosing according to the first distribution of the first probability, and then outputting . The second transition then follows.
Next we have,
where both transitions rely on the fact that . In the last transition, we used the fact that if , thus if then the inequality holds. Else, and the expression on the third line is at most so the inequality on the last transition holds trivially. We will now analyze the last two terms separately. The second term can be bounded as follows,
The first transition is evident, for the second transition note that the numerator does not decrease, while the denominator is the same (it follows from in Equation (37)), finally the third transition uses Equation (38) and the fact that is large from the fourth item of Corollary 8.2.
For the first term, note that,
where we use the fact that there is a one-to-one correspondence between , and . Also, recalling the definition of , we have , so the conditioning in both probabilities is the same, and therefore the two probabilities are equivalent. Next, it is clear that , so using Corollary 8.2, we have,
Putting everything together, we get that
establishing property 6.
Property 4. Take an arbitrary with corresponding such that Equation (39) holds. We have
The first transition is by Equation (39), the second transition is by Equation (37), the third transition is by the fourth item of Corollary 8.2, the fifth transition uses the fact that and , and the last transition uses the fact that . ∎
E.3 Proof of Lemma 8.5
See 8.5
E.3.1 A Necessary Fourier Analytic Fact
We first show a Fourier Analytic fact that will be needed for the proof of Lemma 8.5. Let , for some , and let . We define as follows:
Lemma E.3.
Fix a subspace , then for any we have,
Proof.
If , then for any , we have for all , so the first case follows.
Now suppose , and without loss of generality say that . We can write,
We will show that . Notice that it is sufficient to show that takes each value in with equal probability over uniformly random . First, since , . Next note for any ,
Therefore, takes each of the nonzero values in with probability over uniform , and this concludes the proof. ∎
Lemma E.4.
If has codimension and satisfies,
then there is a nonzero such that and
Proof.
Note that, , so
Using the Fourier decomposition of , we can write,
Using the previous inequality, and the fact that , and Lemma E.3, we have
By the triangle inequality we have,
and finally by the assumption in the lemma statement we have,
Since and there are at most tuples , the result follows. ∎
E.3.2 The Proof of Lemma 8.5
For an arbitrary fixed point , let denote the number of -dimensional subspaces containing . We note that does not depend on which point is fixed. Also let,
(40) |
Also for an arbitrary and , define
(41) |
A straightforward computation shows that , where recall in . We start by removing all that do not satisfy,
(42) |
By Lemma 5.16, the number of removed is at most and is negligible, so for the remainder of the section we assume that all satisfy the above inequalities.
We now define two distributions that we will later show are close to each other. The first is , generated by choosing uniformly and uniformly conditioned on . The second is , generated by choosing uniformly, uniformly, and then uniformly conditioned on . We have
(43) |
By construction of in Lemma 8.4, we have for all . Also, since , the fifth property of Lemma 8.3 yields , for all . Now, noting that , we have:
(44) |
and
(45) |
By construction, for every . Call a triplet bad if
If the triplet is not bad, then by the above inequalities
(46) |
We start by showing that there are very few bad triplets.
Claim E.5.
For each , the number of such that
is at most . Additionally, for every , we have
Proof.
Fix a point , let be the restriction of to the zoom-in of , where if and otherwise. Let . For any satisfying the inequality of the lemma,
We can then apply Lemma E.4 to the zoom-in on . By Lemma E.4, if is bad then there must be such that and,
(47) |
Since by Parseval’s inequality the sum of is at most , there are at most tuples satisfying (47). Now consider a bipartite graph where the left side consists of these tuples , the right side consists of , and the edges are between pairs that satisfy
It follows that the number of edges in this graph is an upper bound on the number of bad triples containing . Since is -generic, we have
for all distinct. Therefore, any two neighbours of a vertex on the left must either have their or be equal, and hence the maximum degree of a vertex on the left side is at most . As a result, the graph has at most edges, where we also use that . This completes the proof of the first assertion of the claim.
For the second part of the lemma, note that by Lemma 8.4. Therefore, for any , we have,
It follows that,
∎
Lemma E.6.
Let be any event defined with respect to . Then,
Proof.
If the triple is not bad, then . Otherwise, we can use the generic bound , which can be obtained from Equation (E.3.2). By the bound on the number of bad triples per in Claim E.5, it follows that
Note that in the last transition we are using the fact that is large by the fourth property in Lemma 8.3. ∎
Now let be the distribution obtained by choosing uniformly, and then choosing uniformly. We have
Using essentially the same proof, we get the following lemma.
Lemma E.7.
Let be any event defined with respect to . Then,
Proof.
We are now ready to prove Lemma 8.5.