This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Limits of CDCL Learning via Merge Resolution

Marc Vinyals    Chunxiao Li    Noah Fleming    Antonina Kolokolova    Vijay Ganesh
Abstract

In their seminal work, Atserias et al. and independently Pipatsrisawat and Darwiche in 2009 showed that CDCL solvers can simulate resolution proofs with polynomial overhead. However, previous work does not address the tightness of the simulation, i.e., the question of how large this overhead needs to be. In this paper, we address this question by focusing on an important property of proofs generated by CDCL solvers that employ standard learning schemes, namely that the derivation of a learned clause has at least one inference where a literal appears in both premises (aka, a merge literal). Specifically, we show that proofs of this kind can simulate resolution proofs with at most a linear overhead, but there also exist formulas where such overhead is necessary or, more precisely, that there exist formulas with resolution proofs of linear length that require quadratic CDCL proofs.

1 Introduction

Over the last two decades, CDCL SAT solvers have had a dramatic impact on many areas of software engineering [CGP+08], security [DVT07, XA05], and AI [BF97]. This is due to their ability to solve very large real-world formulas that contain upwards of millions of variables and clauses [MLM21]. Both theorists and practitioners have expended considerable effort in understanding the CDCL algorithm and the reasons for its unreasonable effectiveness in the context of practical applications. While considerable progress has been made, many questions remain unanswered.

Perhaps the most successful set of tools for understanding the CDCL algorithm come from proof complexity, and a highly influential result is the one that shows that idealized models of CDCL can polynomially simulate the resolution proof system, proved independently by Atserias, Fichte, and Thurley [AFT11], and Pipatsrisawat and Darwiche [PD11], building on initial results by Beame et al. [BKS04] and Hertel et al. [HBPV08]. (See also a recent alternative proof by Beyersdorff and Böhm [BB21].) Such simulation results are very useful because they reassure us that whenever a formula has a short resolution proof then CDCL with the right choice of heuristics can reproduce it.

Recent models make assumptions that are closer to real solvers, but pay for that with a polynomial overhead in the simulation. A series of papers have focused on understanding which of the assumptions are needed for these simulations to hold, often using and/or introducing refinements of resolution along the way. For instance, the question of whether restarts are needed, while still open, has been investigated at length, and the pool resolution [Van05] and RTL [BHJ08] proof systems were devised to capture proofs produced by CDCL solvers that do not restart. The importance of decision heuristics has also been explored recently, with results showing that neither static [MPR20] nor VSIDS-like [Vin20] ordering of variables are enough to simulate resolution in full generality (unless VSIDS scores are periodically erased [LFV+20]). In the case of static ordering, the (semi-)ordered resolution proof system [MPR20] was used to reason about such variants of CDCL solvers.

But even if we stay within the idealized model, it is not clear how efficient CDCL is in simulating resolution. The analysis of Pipatsrisawat and Darwiche gives an O(n4)\mathrm{O}(n^{4}) overhead—that is, if a formula over nn variables has a resolution refutation of length LL, then a CDCL proof with no more than O(n4L)\mathrm{O}(n^{4}L) steps exists. Beyersdorff and Böhm [BB21] improved the overhead to O(n3)\mathrm{O}(n^{3}), but we do not know what the optimal is. Furthermore, to the best of our knowledge, prior to our paper, we did not even know if the overhead can be avoided altogether.

1.1 Learning Schemes in CDCL and Connection with Merges

A common feature of CDCL solvers is the use of 1-empowering learning schemes [PD08, AFT11]: that is, they only learn clauses which enable unit propagations that were not possible before. An example of 1-empowering learning scheme is the popular \lfstyle1UIP learning scheme [MS99]. To model this behavior we build upon a connection between 1-empowerment, and merges [And68], i.e., resolution steps involving clauses with shared literals.

Nearly every CDCL solver nowadays uses the First Unique Implication Point (\lfstyle1UIP) learning scheme, where conflict analysis starts with a clause falsified by the current state of the solver and sequentially resolves it with clauses responsible for unit propagations leading to the conflict, until the clause becomes asserting, i.e., unit immediately upon backjumping.

Descriptions of early implementations of CDCL solvers [MS99, MMZ+01] already remark on the importance of learning an asserting clause, since that nudges the solver towards another part of the search space, and consequently early alternative learning schemes explored learning many kinds of asserting clauses. First observe that conflict analysis can be extended to produce other asserting clauses that appear after the \lfstyle1UIP during conflict analysis such as intermediate UIPs and the last UIP [BS97]. The early solver GRASP can even learn multiple UIP clauses from a single conflict. While there is empirical evidence that it is often best to stop conflict analysis at the \lfstyle1UIP [ZMMM01], recent work has identified conditions where it is advantageous to continue past it [FB20] (see also the discussion of learning schemes therein).

Ryan [Rya04, §2.5] also observed empirically that clause quality is negatively correlated with the length of the conflict analysis derivation and considered the opposite approach, that is, learning clauses that appear before the \lfstyle1UIP during conflict analysis in addition to the \lfstyle1UIP. This approach is claimed to be useful for some empirical benchmarks but, like any scheme that learns multiple clauses, slows down Boolean constraint propagation (BCP) in comparison to a scheme that learns just the \lfstyle1UIP.

Later works provide a more theoretically oriented approach to understanding the strength of \lfstyle1UIP and to learning clauses that appear before the \lfstyle1UIP [DHN07, PD08]. In particular, and highly relevant for our discussion, Pipatsrisawat and Darwiche identified 1-empowerment as a fundamental property of asserting clauses. Furthermore they identified a connection between 1-empowering clauses and merges, and used the simplicity of checking for merges as an approximation for 1-empowerment.

An orthogonal approach is to extend the \lfstyle1UIP derivation by resolving it with clauses other than those that would usually be used during conflict analysis [ABH+08]. A prominent example is clause minimization [SB09], where literals are eliminated from the \lfstyle1UIP clause by resolving it with the appropriate input clauses, independently of their role in the conflict, so the resultant clause that is actually learned is a shorter and therefore stronger version of the \lfstyle1UIP.

Furthermore, a relation between merges and unit-resolution completeness has also been observed in the context of knowledge compilation [dV94]. Finally, the amount of merges directly inferable from a formula (i.e., in a single resolution step) has been proposed, under the name of mergeability, as a measure to help explain the hardness of a formula based on both controlled experiments as well as analysis of real-world instances [ZMW+18].

To summarize, merges are relevant in the context of CDCL learning schemes for the following reason: all practical CDCL learning schemes either produce a 1-empowering clause or extend one, and since 1-empowering clauses always contain a merge in its derivation, we have that all practical learning schemes produce a clause that contains a merge in its derivation, which is exactly the property imposed by the proof systems we introduce below.

1.2 Our Contributions

As mentioned earlier, we build upon a connection between 1-empowerment and merges [PD08, AFT11], and introduce a proof system RMA (for “resolution with merge ancestors”) which includes CDCL with an arbitrary 1-empowering learning scheme. The “merge ancestors” in the name of this system comes from the fact that for any 1-empowering clause, at least one step in its resolution derivation must resolve two clauses that share a common literal: a merge step in the sense of [And68]. Clause minimization procedures, as long as they are applied on top of 1-empowering clauses, are also modelled by RMA.

We prove that, on the one hand, RMA is able to simulate resolution only with a linear overhead. On the other hand, we show a quadratic separation between resolution and RMA, that is there exist formulas with resolution proofs of linear length that require RMA proofs of quadratic length. That is, we show that CDCL may be polynomially worse than resolution because of the properties of a standard learning scheme, but that the blow-up due to these properties is not more than linear.

We also consider weaker proof systems, all of which contain \lfstyle1UIP (and do so with finer granularity), but not necessarily other asserting learning schemes. A technical point of interest is that we work with proof systems that are provably not closed under restrictions, which is unusual in proof complexity. This fact forces our proof to exploit syntactic properties of the proof system, as opposed to relying on more convenient semantic properties.

2 Preliminaries

A literal is either a variable x1=xx^{1}=x or its negation x0=x¯x^{0}=\overline{x}. A clause is a disjunction of literals, and a CNF formula is a conjunction of clauses. The support of a clause or vars(C)\mathrm{vars(C)} is the set of variables it contains. A resolution derivation from a formula FF is a sequence of clauses η=C1,,CL\eta=C_{1},\ldots,C_{L} such that CiC_{i} is either an axiom in FF or it is the conclusion of applying the resolution rule

Res(Ax,Bx¯)=AB\mathrm{Res}(A\lor x,B\lor\overline{x})=A\lor B

on two premises CjC_{j}, CkC_{k} with j,k<ij,k<i. The variable xx that appears with opposite signs in the premises of a resolution inference is called the pivot. If furthermore there is a literal common to AA and BB the resolvent is called a merge. If instead of being the result of a syntactic inference we allow CiC_{i} to be any clause semantically implied by CjC_{j} and CkC_{k}, even if CjC_{j} and CkC_{k} might not be resolvable, then we say η\eta is a semantic resolution derivation. A derivation is a refutation if its last clause is the empty clause \bot. We denote η[a,b]={Ciηi[a,b]}\eta[a,b]=\{C_{i}\in\eta\mid i\in[a,b]\}.

We assume that every clause in a derivation is annotated with the premises it is obtained from, which allows us to treat the proof as a DAG where vertices are clauses and edges point from premises to conclusions. When this DAG is a tree we call a derivation tree-like, and when it is a centipede (i.e., a maximally unbalanced tree) we call it input.

A derivation is unit if in every inference at least one of the premises is a unit clause consisting of a single literal. Since neither input nor unit resolution are complete proof systems, we write FiCF\vdash_{i}C (respectively F1CF\vdash_{1}C) to indicate that there exists an input (resp. unit) resolution derivation of CC from FF.

A clause CC syntactically depends on an axiom AA with respect to a derivation η\eta if there is a path from AA to CC in the DAG representation of η\eta. This does not imply that AA is required to derive CC, since a different derivation might not use AA.

A restriction to variables is a mapping ρ:XX{0,1}\rho\colon X\to X\cup\{0,1\}, successively extended to literals, clauses, formulas, and refutations, simplifying where needed. We write ρ(x)=\rho(x)=* as a shorthand for ρ(x)=x\rho(x)=x. It is well-known that if η\eta is a resolution derivation from FF and ρ\rho is a restriction, then η\upharpoonrightρ{\eta}{\upharpoonright_{\rho}} is a semantic resolution derivation from F\upharpoonrightρ{F}{\upharpoonright_{\rho}}.

It is convenient to leave satisfied clauses in place in a derivation that is the result of applying a restriction to another derivation so that we can use the same indices to refer to both derivations. To do that we use the symbol 11 and treat it as a clause that always evaluates to true, is not supported on any set, does not depend on any clause, and cannot be syntactically resolved with any clause.

A semantic derivation can be turned into a syntactic derivation by ignoring unnecessary clauses. Formally, if η\eta is a semantic resolution derivation, we define its syntactic equivalent s(η)s(\eta) as the following syntactic resolution derivation. Let CηC\in\eta and let AA and BB be the parents of CC. If s(A)\vDashCs(A)\vDash C we set s(C)=s(A)s(C)=s(A), analogously with s(B)s(B). Otherwise we set s(C)=Res(s(A),s(B))s(C)=\mathrm{Res}(s(A),s(B)). It is not hard to see that for each CiηC_{i}\in\eta, s(Ci)\vDashCis(C_{i})\vDash C_{i}.

2.1 CDCL

We need to define a few concepts from CDCL proofs. An in-depth treatment can be found in the Handbook of Satisfiability [BN21]. Fix a CNF FF, also known as clause database. A trail τ\tau is a sequence of tuples (xji=b,Ci)(x_{j_{i}}=b,C_{i}) where CiC_{i} is either a clause in FF or the special symbol dd representing a decision. We denote by α<i\alpha_{<i} the assignment {xji=bi<i}\{x_{j_{i}}=b\mid i^{\prime}<i\}, and we denote by dl(i)=dl(i1)+Ci=d\mathrm{dl}(i)=\mathrm{dl}(i-1)+\llbracket{C_{i}=d}\rrbracket the decision level at position ii, that is the number of decisions up to ii. We mark the position of the last decision in a trail by ii^{*}.

A trail is valid if for every position ii that is not a decision we have that Ci\upharpoonrightα<i=xjib{C_{i}}{\upharpoonright_{\alpha_{<i}}}=x_{j_{i}}^{b} and for every decision ii we have that for every clause CFC\in F such that C\upharpoonrightα<i=xb{C}{\upharpoonright_{\alpha_{<i}}}=x^{b}, the literal xbx^{b} appears in the trail before ii. In particular, for every position i<ii^{\prime}<i with dl(i)<dl(i)\mathrm{dl}(i^{\prime})<\mathrm{dl}(i) we have Ci\upharpoonrightα<ixjib{C_{i}}{\upharpoonright_{\alpha_{<i^{\prime}}}}\neq x_{j_{i}}^{b}.

A clause CC is asserting if it is unit at the last decision in the trail, that is C\upharpoonrightα<i=xb{C}{\upharpoonright_{\alpha_{<i^{*}}}}=x^{b}. It is 1-empowering if CC is implied by FF and can lead to new unit propagations after being added to FF, that is if there exists a literal C\ell\in C such that for some A{,}A\in\{\bot,\ell\}, it holds that FC¯\nvdash1AF\land\overline{C\setminus\ell}\nvdash_{1}A. If a clause is not 1-empowering then we say it is absorbed by FF.

Given a clause D|τ|D_{\lvert\tau\rvert} falsified by a trail τ\tau, the conflict derivation is an input derivation D|τ|,,Dk+1,DkD_{\lvert\tau\rvert},\ldots,D_{k+1},D_{k} where Di1=Res(Di,Ci)D_{i-1}=\mathrm{Res}(D_{i},C_{i}) if xjiDix_{j_{i}}\in D_{i}, and Di1=DiD_{i-1}=D_{i} otherwise. The first (i.e., with the largest index) asserting clause in the derivation is called the \lfstyle1UIP. Note that DiD_{i^{*}} is always asserting (because DiD_{i} is falsified by αi\alpha_{\leq i} for ii|τ|i^{*}\leq i\leq\lvert\tau\rvert and DiD_{i^{*}} is not falsified by α<i\alpha_{<i^{*}}), therefore we can assume that the \lfstyle1UIP always has index at least ii^{*}.

We call a sequence of input derivations input-structured if the last clause of each derivation can be used as an axiom in successive derivations. The last clause of each but the last derivation is called a lemma. A CDCL derivation is an input-structured sequence of conflict derivations, where learned clauses are lemmas. This definition is similar to that of Resolution Trees with Input Lemmas [BHJ08], with the difference that the sequence only needs to be ordered, without imposing any further tree-structure on the global proof.

The following Lemmas highlight the practical relevance of merges by relating them to \lfstyle1UIP, asserting, and 1-empowering clauses.

Lemma 2.1 ([PD08, Proposition 2]).

If a clause is asserting, then it is 1-empowering.111The original result does not prove 1-consistency, but the proof is analogous.

Lemma 2.2 ([AFT11, Lemma 8]).

If AxA\lor x and Bx¯B\lor\overline{x} are absorbed but ABA\lor B is 1-empowering, then ABA\lor B is a merge. In particular, if a clause is 1-empowering, then it contains a merge in its derivation.

Lemma 2.3.

The \lfstyle1UIP clause is a merge.

Proof.

Let Dj=Res(Cj+1,Dj+1)D_{j}=\mathrm{Res}(C_{j+1},D_{j+1}) be the \lfstyle1UIP. On the one hand, since every clause in the trail contains at least two literals at the same decision level it appears in, Cj+1C_{j+1} contains two literals at the last decision level. On the other hand, any clause that is not in the trail also contains two literals at the last decision level, and in particular D|τ|D_{\lvert\tau\rvert}. Since |Di+1Di|1\lvert D_{i+1}\setminus D_{i}\rvert\leq 1 and Dj+1D_{j+1} is not asserting, it also contains two literals at the last decision level.

We accounted for 4 literals at the last decision level present in the premises of DjD_{j}, of which 2 are not present in the conclusion because they are the pivots. In order for DjD_{j} to contain only one literal at the last decision level, the remaining two literals must be equal. ∎

3 Proof Systems

We define our proof systems in terms of the input-structured framework. Every resolution proof can be thought of as being input-structured if we consider it as a sequence of unit-length input resolutions and every clause as a lemma; it is when we impose restrictions on which clauses are permitted as lemmas that we obtain different proof systems. The diagram in Figure 1 can help keeping track of the proof systems.

ResfgfgRMAfgfgLRMAfgfgRELfgfgRMLfgfgLRMLfgfgLREMLfgfg\lfstyle1UIPfgfgAssertingfgfg5.1fgfg6.9fgfg6.1fgfg6.5fgfg6.5fgfg6.9fgfg
Figure 1: Relations between proof systems. A solid arrow AAfgfgBBfgfg indicates that AA simulates BB with no overhead. A dashed arrow AAfgfgBBfgfg indicates that AA simulates BB with no overhead, but BB requires linear overhead to simulate AA. Statements proving separations are referenced.

Andrews’ definition of merge resolution [And68] considers tree-like proofs with the additional restriction that in every inference at least one premise is an axiom or a merge. He also observes that such derivations can be made input-structured.

Observation 3.1 ([And68]).

A tree-like merge resolution derivation can be decomposed into an input-structured sequence where all the lemmas are merges.

This observation is key when working with such derivations, as is apparent in Sections 4 and A, to the point that we use as an alternative way to define merge resolution.

Andrews’ main result is that the merge restriction does not affect tree-like resolution.

Lemma 3.2 ([And68, Lemma 5]).

If there is a tree-like resolution derivation of CC of length LL where at most the root is a merge, then there is an input resolution derivation of some CCC^{\prime}\subseteq C of length at most LL.

Theorem 3.3 ([And68, Theorem 1]).

If there is a tree-like resolution derivation of CC of length LL, then there is a tree-like merge resolution derivation of some CCC^{\prime}\subseteq C of length at most LL.

If we lift the tree-like restriction from the input-structured view of merge resolution proofs we obtain a proof system between tree- and DAG-like resolution where clauses can be reused (i.e., have outdegree larger than 11) if and only if they are merges or, in other words, lemmas in the input-structured decomposition. We call this proof system Resolution with Merge Lemmas and refer to it with the acronym RML.

Definition 3.4.

A RML derivation is an input-structured sequence of unit resolution derivations where all lemmas are merges.

CDCL refutations produced by solvers that use the \lfstyle1UIP learning scheme are in RML form, as a consequence of Lemma 2.3. We can also generalize RML to allow reusing clauses that contain a merge anywhere in their derivation. We call this proof system Resolution with Merge Ancestors, or RMA for short.

Definition 3.5.

A RMA derivation is an input-structured sequence of unit resolution derivations where all derivations but the last contain a merge.

Note that by Lemma 3.2 it does not matter if we require the sequence of derivations of an RMA derivation to be input derivations or if we allow general trees. In fact, our lower bound results hold for a more general proof system where we only ask that every clause with outdegree larger than 11 has an ancestor that is a merge. Such proof system does not have a simple input structure, but can rather be thought of as a sequence of tree-like resolution derivations whose roots are merges, followed by a standard resolution derivation using the roots of the previous derivations as axioms.

To make the connection back to CDCL, we can define a proof system called Resolution with Empowering Lemmas that captures CDCL refutations produced by solvers that use any asserting learning scheme or 1-empowering learning scheme.

Definition 3.6.

Let C1,,CL1C_{1},\ldots,C_{L-1} be the lemmas of an input-structured sequence of unit derivations. The sequence is a Resolution with Empowering Lemmas (REL) derivation of a formula FF if CiC_{i} is 1-empowering with respect to F{Cj:j<i}F\cup\{C_{j}:j<i\} for all i[1,L1]i\in[1,L-1].

It follows from Lemmas 2.1 and 2.2 that such refutations are in RMA form.

Observation 3.7.

A REL derivation is a RMA derivation.

It might seem more natural to work with the REL proof system rather than its merge-based counterparts, since REL is defined exactly through the 1-empowering property. However, while the merge property is easy to check because it is local to the derivation at hand, we can only determine if a clause is 1-empowering by looking at the full history of the derivation, in particular what the previous lemmas are. This makes REL too cumbersome to analyse. Furthermore, CDCL refutations produced apply a clause minimization scheme on top of an asserting clause might not be in REL form, but they are still in RMA form.

A further property of input derivations produced by a CDCL solver is that once a variable is resolved, it does not appear later in the derivation.

Definition 3.8.

A resolution derivation η\eta is strongly regular if for every resolution step ii, the pivot variable xix_{i} is not part of the support of any clause Ciη[i,L]C_{i}\in\eta[i,L]. A sequence of derivations is locally regular if every derivation in the sequence is strongly regular. A LRML derivation (resp. LRMA) is a locally regular RML derivation (resp. RMA).

Finally we can consider derivations that have empowering, merge lemmas and are locally regular. These still include \lfstyle1UIP proofs.

Definition 3.9.

A LREML derivation is a derivation that is both LRML and REL.

It follows from the simulation of resolution by CDCL [PD11, AFT11] that all (DAG-like) proof systems we defined polynomially simulate standard resolution. In Section 4 we make this simulation more precise and prove that the simulation overhead can be made linear, and in Section 5 that the simulation is optimal because there exist formulas that have resolution refutations of linear length but require RMA refutations of quadratic length.

4 Simulation

As an auxiliary tool to simulate resolution in RML we define the input-resolution closure of a set DD, denoted Cli(D)={CCC,DiC}\mathrm{Cl_{i}}(D)=\{C\mid\exists C^{\prime}\subseteq C,\,D\vdash_{i}C^{\prime}\}, as the set of clauses derivable from DD via input resolution plus weakening. It is well-known that, since input resolution derivations can be assumed to be strongly regular without loss of generality, we can also assume them to be at most linear in the number of variables.

Observation 4.1.

If DD is a CNF formula over nn variables and CCli(D)C\in\mathrm{Cl_{i}}(D) then there is a strongly regular input resolution derivation of some CCC^{\prime}\subseteq C from DD of length at most nn.

Combining Theorem 3.3 with the idea that in order to simulate a resolution derivation we do not need to generate each clause, but only do enough work so that in the following steps we can pretend that we had derived it [PD11, AFT11], we can prove that merge resolution simulates resolution with at most a multiplicative linear overhead in the number of variables.

Theorem 4.2.

If FF is a CNF formula over nn variables that has a resolution refutation of length LL then it has a RML refutation of length O(nL)\mathrm{O}(nL).

Proof.

Let π=(C1,,CL)\pi=(C_{1},\ldots,C_{L}) be a resolution refutation. We construct a sequence of sets D0,,DLD_{0},\ldots,D_{L} with the following properties.

  1. 1.

    DtFD_{t}\setminus F is the set of lemmas in a RML derivation of length at most (2n+1)t(2n+1)t.

  2. 2.

    π[1,t]Cli(Dt)\pi[1,t]\subseteq\mathrm{Cl_{i}}(D_{t}).

This is enough to prove the theorem: since Dt\bot\in D_{t} we can obtain \bot from DtD_{t} in length nn, so the total length of the refutation is (2n+1)L+n(2n+1)L+n.

We build the sets by induction, starting with D0=FD_{0}=F. Assume we have built DtD_{t} and let C=Ct+1=Res(A,B)C=C_{t+1}=\mathrm{Res}(A,B) with A,Bπ[1,t]A,B\in\pi[1,t]. If CCli(Dt)C\in\mathrm{Cl_{i}}(D_{t}) we set Dt+1=DtD_{t+1}=D_{t} and we are done. Otherwise, by induction we have A,BCli(Dt)A,B\in\mathrm{Cl_{i}}(D_{t}), therefore by Observation 4.1 there are input resolution derivations of AAA^{\prime}\subseteq A and BBB^{\prime}\subseteq B of length at most nn. Since neither A\vDashCA^{\prime}\vDash C nor B\vDashCB^{\prime}\vDash C, AA^{\prime} and BB^{\prime} can be resolved and therefore there is a tree-like derivation η\eta of CCC^{\prime}\subseteq C from DtD_{t} of length at most 2n+12n+1. By Theorem 3.3 there is a tree-like merge resolution derivation η\eta^{\prime} of C′′CC^{\prime\prime}\subseteq C from DtD_{t} of length at most 2n+12n+1. By Observation 3.1 the derivation η\eta^{\prime} can be decomposed into a sequence of input derivations of total length at most 2n+12n+1. Let EE be the lemmas in that sequence and set Dt+1=DtED_{t+1}=D_{t}\cup E. We have that CCli(FE)Cli(Dt+1)C\in\mathrm{Cl_{i}}(F\cup E)\subseteq\mathrm{Cl_{i}}(D_{t+1}), and that we can obtain EE from DtD_{t} in at most 2n+12n+1 steps. Thus Dt+1D_{t+1} has all the required properties. ∎

We can be a bit more precise with the description of the simulation if we look at the structure of η\eta before applying Theorem 3.3. Let AMA_{M} and BMB_{M} be the last merges in the input derivation of AA^{\prime} and BB^{\prime} respectively, and let E={AM,BM}E=\{A_{M},B_{M}\}.

Now consider the fragment of the input derivation of AA^{\prime} from AMA_{M} to AA^{\prime}, analogously with BB^{\prime}. We have a tree-like derivation of CC^{\prime} where at most the root is a merge, therefore we can apply Lemma 3.2 directly instead of Theorem 3.3 and obtain an input resolution derivation of C′′CC^{\prime\prime}\subseteq C from EFE\cup F.

If we also make sure that the input derivations of AA^{\prime} and BB^{\prime} are strongly regular, we have that LRML can also simulate resolution with the same O(n)\mathrm{O}(n) overhead as RML.

An analogous result can be obtained for LREML from the following lemma.

Lemma 4.3 ([PD11]).

If FF absorbs AxA\lor x and Bx¯B\lor\overline{x}, then FiCABF\vdash_{i}C^{\prime}\subseteq A\lor B.

Corollary 4.4.

If FF is a CNF formula over nn variables that has a resolution refutation of length LL then it has a LREML refutation of length O(nL)\mathrm{O}(nL).

Proof.

The proof follows the general structure of Theorem 4.2, except that we use a sequence of steps DtjD_{t}^{j} in order to construct DtD_{t}. Our induction hypothesis is that DtjD_{t}^{j} can be derived from DtD_{t} in pp inference steps in LREML, and that AA^{\prime} and BB^{\prime} can be derived from DtjD_{t}^{j} in qq steps, with p+q2np+q\leq 2n.

The base case Dt0=DtD_{t}^{0}=D_{t} is trivial.

For the inductive case, assume that the input derivations leading to AA^{\prime} and BB^{\prime} are strongly regular without loss of generality. By Lemma 4.3 either AA^{\prime} or BB^{\prime} is 1-empowering, say AA^{\prime}. Let CC be the first 1-empowering clause in the derivation of AA^{\prime}. By Lemma 2.2 CC is a merge, therefore we can take Dtj+1=Dtj{C}D_{t}^{j+1}=D_{t}^{j}\cup\{C\}. ∎

5 Separation

We prove the following separation between standard resolution and RMA.

Theorem 5.1.

There exists a family of formulas FnF_{n} over O(nlogn)\mathrm{O}(n\log n) variables and O(nlogn)\mathrm{O}(n\log n) clauses that have resolution refutations of length O(nlogn)\mathrm{O}(n\log n) but every RMA refutation requires length Ω(n2logn)\Omega(n^{2}\log n).

5.1 Formula

Let ,m,n\ell,m,n be positive integers. We have variables xix_{i} for i[m1]i\in[m\ell-1] and wj,kw_{j,k} for j[]j\in[\ell] and k[n]k\in[n]. For convenience we define x0=1x_{0}=1 and xm=0x_{m\ell}=0, which are not variables. Let X={xii[m1]}X=\{x_{i}\mid i\in[m\ell-1]\}, Wj={wj,kk[n]}W_{j}=\{w_{j,k}\mid k\in[n]\} and W=j[]WjW=\bigcup_{j\in[\ell]}W_{j}. For each j[]j\in[\ell] we build the following gadget:

wj,k=wj,k+1\displaystyle w_{j,k}=w_{j,k+1} for k[n1]k\in[n-1] (1)

Each equality is expanded into the two clauses Bj,k,1=wj,kwj,k+1¯B_{j,k,1}=w_{j,k}\lor\overline{w_{j,k+1}} and Bj,k,0=wj,k¯wj,k+1B_{j,k,0}=\overline{w_{j,k}}\lor w_{j,k+1}, and we collectively call them 𝒲={Bj,k,bj[],k[n1],b{0,1}}\mathcal{W}=\{B_{j,k,b}\mid j\in[\ell],k\in[n-1],b\in\{0,1\}\}. Observe that the jj-th gadget implies wj,1=wj,nw_{j,1}=w_{j,n}. Additionally we build the following gadget:

(w1,1=w1,n)x1\displaystyle(w_{1,1}=w_{1,n})\rightarrow x_{1} (2)
(wı^,1=wı^,n)(xi1xi)\displaystyle(w_{\hat{\imath},1}=w_{\hat{\imath},n})\rightarrow(x_{i-1}\rightarrow x_{i}) for i[2,m1]i\in[2,m\ell-1] (3)
(w,1=w,n)xm1¯\displaystyle(w_{\ell,1}=w_{\ell,n})\rightarrow\overline{x_{m\ell-1}} (4)

where ı^[]\hat{\imath}\in[\ell] denotes the canonical form of i(mod)i\pmod{\ell}. Each constraint is expanded into the two clauses Ai,1=wı^,1wı^,nxi1¯xiA_{i,1}=w_{\hat{\imath},1}\lor w_{\hat{\imath},n}\lor\overline{x_{i-1}}\lor x_{i} and Ai,0=wı^,1¯wı^,n¯xi1¯xiA_{i,0}=\overline{w_{\hat{\imath},1}}\lor\overline{w_{\hat{\imath},n}}\lor\overline{x_{i-1}}\lor x_{i}, and we collectively call them 𝒳={Ai,bi[m],b{0,1}}\mathcal{X}=\{A_{i,b}\mid i\in[m\ell],b\in\{0,1\}\}. The resulting formula is called F,m,nF_{\ell,m,n}.

5.2 Upper Bound

It is not hard to see that there is a resolution refutation of F,m,nF_{\ell,m,n} of length O((m+n))\mathrm{O}(\ell\cdot(m+n)). Indeed, we first derive the two clauses representing wj,1=wj,nw_{j,1}=w_{j,n} for each j[]j\in[\ell], which requires O(n)\mathrm{O}(n\ell) steps:

  wj,1wj,2¯w_{j,1}\lor\overline{w_{j,2}}         wj,2wj,3¯w_{j,2}\lor\overline{w_{j,3}}             wj,1wj,3¯w_{j,1}\lor\overline{w_{j,3}}                       \vdots                     wj,1wj,n1¯w_{j,1}\lor\overline{w_{j,n-1}}        wj,n1wj,n¯w_{j,n-1}\lor\overline{w_{j,n}}                                wj,1wj,n¯w_{j,1}\lor\overline{w_{j,n}} (5)

Then we resolve each of the 𝒳\mathcal{X} axioms with one of these clauses, appropriately chosen so that we obtain pairs of clauses of the form wı^bxi1¯xiw_{\hat{\imath}}^{b}\lor\overline{x_{i-1}}\lor x_{i} for i[m]i\in[m\ell], and resolve each pair to obtain the chain of implications x1,,xixi+1,,xn1¯x_{1},\ldots,x_{i}\rightarrow x_{i+1},\ldots,\overline{x_{n\ell-1}} in O(m)\mathrm{O}(m\ell) steps.

  wı^,1wı^,n¯w_{\hat{\imath},1}\lor\overline{w_{\hat{\imath},n}}         wı^,1wı^,nxi1¯xiw_{\hat{\imath},1}\lor w_{\hat{\imath},n}\lor\overline{x_{i-1}}\lor x_{i}                 wı^,1xi1¯xiw_{\hat{\imath},1}\lor\overline{x_{i-1}}\lor x_{i}         wı^,1¯wı^,n\overline{w_{\hat{\imath},1}}\lor w_{\hat{\imath},n}         wı^,1¯wı^,n¯xi1¯xi\overline{w_{\hat{\imath},1}}\lor\overline{w_{\hat{\imath},n}}\lor\overline{x_{i-1}}\lor x_{i}               wı^,1¯xi1¯xi\overline{w_{\hat{\imath},1}}\lor\overline{x_{i-1}}\lor x_{i}                                                   xi1¯xi\overline{x_{i-1}}\lor x_{i} (6)

Since we have derived a chain of implications x1x_{1}, x1x2x_{1}\rightarrow x_{2}, …, xm1xm1x_{m\ell-1}\rightarrow x_{m\ell-1}, xm1¯\overline{x_{m\ell-1}} we can complete the refutation in O(m)\mathrm{O}(m\ell) more steps. Let us record our discussion.

Lemma 5.2.

F,m,nF_{\ell,m,n} has a resolution refutation of length O((m+n))\mathrm{O}(\ell\cdot(m+n)).

Before we prove the lower bound let us discuss informally what are the natural ways to refute this formula in RML, so that we understand which behaviours we need to rule out.

If we try to reproduce the previous resolution refutation, since we cannot reuse the clauses representing wj,1=wj,nw_{j,1}=w_{j,n} because they are not merges, we have to rederive them each time we need them, which means that it takes O(mn)\mathrm{O}(mn\ell) steps to derive the chain of implications x1,,xixi+1,,xn1¯x_{1},\ldots,x_{i}\rightarrow x_{i+1},\ldots,\overline{x_{n\ell-1}}. We call this approach refutation 1. This refutation has merges (over wı^,1w_{\hat{\imath},1}, xi1x_{i-1}, and xix_{i}) when we produce wı^,1bxi1¯xiw_{\hat{\imath},1}^{b}\lor\overline{x_{i-1}}\lor x_{i}, and (over xi1x_{i-1} and xix_{i}) when we produce xi1¯xi\overline{x_{i-1}}\lor x_{i}, but since we never reuse these clauses the refutation is in fact tree-like.

An alternative approach, which we call refutation 2, is to start working with the 𝒳\mathcal{X} axioms instead. In this proof we clump together all of the repeated constraints of the form wj,1wj,nw_{j,1}\neq w_{j,n} for every j[]j\in[\ell], and then resolve them out in one go. In other words, we first derive the sequence of constraints

Di=(ı^[min(i,)]wı^,1wı^,n)xi\displaystyle D_{i}=\biggl{(}\bigvee_{\hat{\imath}\in[\min(i,\ell)]}w_{\hat{\imath},1}\neq w_{\hat{\imath},n}\biggr{)}\lor x_{i} for i[m],\displaystyle\text{for $i\in[m\ell]$}\enspace, (7)

where DiD_{i} can be obtained from Di1D_{i-1} and the pair of 𝒳\mathcal{X} axioms Ai,bA_{i,b}, then resolve away the inequalities from Dm=j[]wj,1wj,nD_{m\ell}=\bigvee_{j\in[\ell]}w_{j,1}\neq w_{j,n} using the 𝒲\mathcal{W} axioms. However, representing any of the constraints DiD_{i} for ii\geq\ell requires 22^{\ell} clauses, which is significantly larger than mnmn\ell and even superpolynomial for large enough \ell, so this refutation is not efficient either. Note that this refutation has merges (over WW variables) each time that we derive DiD_{i} with ii\geq\ell.

A third and somewhat contrived way to build a refutation is to derive the pair of clauses representing wj,1=wj,nw_{j,1}=w_{j,n} using a derivation whose last step is a merge, so that they can be reused. Each of these clauses can be derived individually in O(mn)\mathrm{O}(mn\ell) steps, for a total of O(mn2)\mathrm{O}(mn\ell^{2}) steps, by slightly adapting refutation 5.2, substituting each derivation of xixi+1x_{i}\rightarrow x_{i+1} by a derivation of wj,1wj,n¯xi¯xi+1w_{j,1}\lor\overline{w_{j,n}}\lor\overline{x_{i}}\lor x_{i+1} whenever ij(mod)i\equiv j\pmod{\ell} so that at the end we obtain wj,1wj,n¯w_{j,1}\lor\overline{w_{j,n}} instead of the empty clause. Such a substitution clause can be obtained, e.g., by resolving wj,1wj,2xi¯xi+1w_{j,1}\lor w_{j,2}\lor\overline{x_{i}}\lor x_{i+1} with wj,2¯wj,n¯xi¯xi+1\overline{w_{j,2}}\lor\overline{w_{j,n}}\lor\overline{x_{i}}\lor x_{i+1} as follows

wj,2wj,3¯w_{j,2}\lor\overline{w_{j,3}} wj,3wj,4¯w_{j,3}\lor\overline{w_{j,4}}        wj,2wj,4¯w_{j,2}\lor\overline{w_{j,4}}                \vdots             wj,2wj,n1¯w_{j,2}\lor\overline{w_{j,n-1}}  wj,n1wj,n¯w_{j,n-1}\lor\overline{w_{j,n}}                     wj,2wj,n¯w_{j,2}\lor\overline{w_{j,n}}  wı^,1wı^,nxi1¯xiw_{\hat{\imath},1}\lor w_{\hat{\imath},n}\lor\overline{x_{i-1}}\lor x_{i}                                      wı^,1wı^,2xi1¯xiw_{\hat{\imath},1}\lor w_{\hat{\imath},2}\lor\overline{x_{i-1}}\lor x_{i}   wı^,1wı^,2¯w_{\hat{\imath},1}\lor\overline{w_{\hat{\imath},2}} wı^,1¯wı^,n¯xi1¯xi\overline{w_{\hat{\imath},1}}\lor\overline{w_{\hat{\imath},n}}\lor\overline{x_{i-1}}\lor x_{i}         wı^,2¯wı^,n¯xi1¯xi\overline{w_{\hat{\imath},2}}\lor\overline{w_{\hat{\imath},n}}\lor\overline{x_{i-1}}\lor x_{i}                                                               wı^,1wı^,n¯xi1¯xiw_{\hat{\imath},1}\lor\overline{w_{\hat{\imath},n}}\lor\overline{x_{i-1}}\lor x_{i} (8)

After deriving wj,1=wj,nw_{j,1}=w_{j,n} as merges we follow the next steps of refutation 5.2 and complete the refutation in O(m)\mathrm{O}(m\ell) steps. We call this refutation 3.

Observe that the minimum length of deriving the clauses representing wj,1=wj,nw_{j,1}=w_{j,n} is only O(n)\mathrm{O}(n), even in RML, so if we only used the information that refutation 5.2 contains these clauses we would only be able to bound its length by Ω((m+n))\Omega(\ell\cdot(m+n)). Therefore when we compute the hardness of deriving a clause we need to take into account not only its semantics but how it was obtained syntactically.

5.3 Lower Bound

Before we begin proving our lower bound in earnest we make two useful observations.

Lemma 5.3.

Let η\eta be a resolution derivation that only depends on the 𝒲\mathcal{W} axioms. Then η\eta does not contain any merges, and all clauses are supported on WW.

Proof.

We prove by induction that every clause in η\eta is of the form wj,kwj,k¯w_{j,k}\lor\overline{w_{j,k^{\prime}}} with kkk\neq k^{\prime}. This is true for the axioms. By induction hypothesis, a generic resolution step over wj,kw_{j,k} is of the form

  wj,kwj,k¯w_{j,k}\lor\overline{w_{j,k^{\prime}}}         wj,k¯wj,k′′\overline{w_{j,k}}\lor w_{j,k^{\prime\prime}}             wj,k′′wj,k¯w_{j,k^{\prime\prime}}\lor\overline{w_{j,k^{\prime}}} (9)

and in particular is not a merge. ∎

Lemma 5.4.

Let η\eta be a resolution derivation of a clause CC supported on WW variables that uses an 𝒳\mathcal{X} axiom. Then η\eta uses at least one Ai,bA_{i,b} axiom for each i[m]i\in[m\ell].

Proof.

We prove the contrapositive and assume that there is an axiom Ai,bA_{i,b} that is used, and either both Ai+1,0A_{i+1,0} and Ai+1,1A_{i+1,1} are not used, or both Ai1,0A_{i-1,0} and Ai1,1A_{i-1,1} are not. In the first case the literal xix_{i} appears in every clause in the path from Ai,bA_{i,b} to CC, contradicting that CC is supported on WW variables. Analogously with literal xi1¯\overline{x_{i-1}} in the second case. ∎

Our first step towards proving the lower bound is to rule out that refutations like refutation 5.2 can be small, and to do that we show that wide clauses allow for very little progress. This is a common theme in proof complexity, and the standard tool is to apply a random restriction to a short refutation in order to obtain a narrow refutation. However, RMA is not closed under restrictions, as we prove later in Corollary 5.12, and because of this we need to argue separately about which merges are preserved.

Let us define the class of restrictions that we use and which need to respect the structure of the formula. A restriction is an autarky [MS85] with respect to a set of clauses DD if it satisfies every clause that it touches; in other words for every clause CDC\in D either C\upharpoonrightρ=1{C}{\upharpoonright_{\rho}}=1 or C\upharpoonrightρ=C{C}{\upharpoonright_{\rho}}=C. A restriction is kk-respecting if it is an autarky with respect to 𝒲\mathcal{W} axioms, we have F,m,n\upharpoonrightρFk,m,n{F_{\ell,m,n}}{\upharpoonright_{\rho}}\cong F_{k,m,n} up to variable renaming, and every XX variable is mapped to an XX variable. Our definition of a narrow clause is also tailored to the formula at hand, and counts the number of different WW-blocks that a clause CC mentions. Formally μ(C)=|{j[]xj,kvars(C)}|\mu(C)=\lvert\{j\in[\ell]\mid\exists x_{j,k}\in\mathrm{vars(C)}\}\rvert.

Lemma 5.5.

Let π\pi be a resolution refutation of F,m,nF_{\ell,m,n} of length L=o((4/3)/8)L=\mathrm{o}((4/3)^{\ell/8}). There exists an /4\ell/4-respecting restriction ρ\rho such that every clause in π\upharpoonrightρ{\pi}{\upharpoonright_{\rho}} has μ(C)/8\mu(C)\leq\ell/8.

Proof.

We use the probabilistic method. Consider the following distribution 𝒥\mathcal{J} over {0,1,}\{0,1,*\}^{\ell}: each coordinate is chosen independently with Pr[Ji=0]=Pr[Ji=1]=1/4\Pr[J_{i}=0]=\Pr[J_{i}=1]=1/4, Pr[Ji=]=1/2\Pr[J_{i}=*]=1/2. Given a random variable J𝒥J\sim\mathcal{J} sampled according to this distribution, we derive a random restriction ρ\rho as follows: ρ(wj,i)=Jj\rho(w_{j,i})=J_{j}, ρ(xi)=\rho(x_{i})=* if Jı^=J_{\hat{\imath}}=*, and ρ(xi)=ρ(xi1)\rho(x_{i})=\rho(x_{i-1}) otherwise (where ρ(x0)=1\rho(x_{0})=1).

Observe that F,m,n\upharpoonrightρF|J1()|,m,n{F_{\ell,m,n}}{\upharpoonright_{\rho}}\cong F_{\lvert J^{-1}(*)\rvert,m,n} up to variable renaming, and by a Chernoff bound we have Pr[|J1()|</4]e/16\Pr[\lvert J^{-1}(*)\rvert<\ell/4]\leq e^{-\ell/16}.

We also have, for every clause CπC\in\pi with μ(C)>/8\mu(C)>\ell/8, that

Pr[C\upharpoonrightρ1](3/4)μ(C)(3/4)/8.\Pr[{C}{\upharpoonright_{\rho}}\neq 1]\leq(3/4)^{\mu(C)}\leq(3/4)^{\ell/8}\enspace. (10)

Therefore by a union bound the probability that |J1()|</4\lvert J^{-1}(*)\rvert<\ell/4 or that any clause has μ(C\upharpoonrightρ)>/8\mu({C}{\upharpoonright_{\rho}})>\ell/8 is bounded away from 11 and we conclude that there exists a restriction ρ\rho that satisfies the conclusion of the lemma. ∎

Note that s(π\upharpoonrightρ)s({\pi}{\upharpoonright_{\rho}}) is a resolution refutation of Fn,\upharpoonrightρ{F_{n,\ell}}{\upharpoonright_{\rho}}, but not necessarily a RMA refutation, therefore we lose control over which clauses may be reused222Recall that s(π)s(\pi) is the syntactic equivalent of π\pi.. Nevertheless, we can identify a fragment of s(π\upharpoonrightρ)s({\pi}{\upharpoonright_{\rho}}) where we still have enough information.

Lemma 5.6.

There exists an integer tt such that ψ=s(π[1,t]\upharpoonrightρ)\psi=s({\pi[1,t]}{\upharpoonright_{\rho}}) is a resolution derivation of a clause supported on WW variables that depends on an 𝒳\mathcal{X} axiom and where no clause supported on WW variables is reused.

Proof.

Let CtπC_{t}\in\pi be the first clause that depends on an 𝒳\mathcal{X} axiom and such that Dt=s(Ct\upharpoonrightρ)D_{t}=s({C_{t}}{\upharpoonright_{\rho}}) is supported on WW, which exists because \bot is one such clause.

By definition of tt, we have that every ancestor DkψD_{k}\in\psi of DtD_{t} that is supported on WW variables corresponds to a clause CkC_{k} in π\pi that only depends on 𝒲\mathcal{W} axioms, hence by Lemma 5.3 CkC_{k} is not a merge. By definition of RMA CkC_{k} is not reused, and by construction of s()s(\cdot) neither is DkD_{k}.

It remains to prove that DtD_{t} depends on an 𝒳\mathcal{X} axiom. Since CtC_{t} depends on an 𝒳\mathcal{X} axiom, at least one of its predecessors CpC_{p} and CqC_{q} also does, say CpC_{p}. By definition of tt, Dp=s(Cp\upharpoonrightρ)D_{p}=s({C_{p}}{\upharpoonright_{\rho}}) is not supported on WW, and hence by Lemma 5.3 either DpD_{p} depends on an 𝒳\mathcal{X} axiom or Dp=1D_{p}=1. Analogously, if CqC_{q} also depends on an 𝒳\mathcal{X} axiom then so does Dq=s(Cj\upharpoonrightρ)D_{q}=s({C_{j}}{\upharpoonright_{\rho}}) (or it is 11) and we are done. Otherwise CqC_{q} is of the form wj,kwj,k¯w_{j,k}\lor\overline{w_{j,k^{\prime}}} and is either satisfied by ρ\rho or left untouched. In both cases we have that Dq\vDashCt\upharpoonrightρD_{q}\not\vDash{C_{t}}{\upharpoonright_{\rho}} (trivially in the first case and because DqD_{q} contains the pivot while CtC_{t} does not in the second), hence DtD_{t} depends on DpD_{p}. ∎

Note that CtC_{t} may be semantically implied by the 𝒲\mathcal{W} axioms, and have a short derivation as in refutation 5.2, therefore we are forced to use syntactic arguments to argue that deriving CtC_{t} using an 𝒳\mathcal{X} axiom takes many resolution steps.

The next step is to break ψ\psi into mm (possibly intersecting) parts, each corresponding roughly to the part of ψ\psi that uses 𝒳\mathcal{X} axioms with variables in an interval of length \ell (by Lemma 5.4 we can assume that ψ\psi contains axioms from every interval). To do this we use the following family of restrictions defined for i[n]i\in[n]:

σi(xi)\displaystyle\sigma_{i}(x_{i^{\prime}}) ={1if iiif i<i(i+1)0if (i+1)<i\displaystyle=\begin{cases*}1&if $i^{\prime}\leq i\ell$\\ *&if $i\ell<i^{\prime}\leq(i+1)\ell$\\ 0&if $(i+1)\ell<i^{\prime}$\end{cases*} σi(wi,j)\displaystyle\sigma_{i}(w_{i^{\prime},j}) =\displaystyle=* (11)

Let Xi=Xσi1()X_{i}=X\cap\sigma_{i}^{-1}(*) and note that F,m,n\upharpoonrightσiF,1,n{F_{\ell,m,n}}{\upharpoonright_{\sigma_{i}}}\cong F_{\ell,1,n}.

Clauses in ψ\psi with many XX variables could be tricky to classify, but intuitively it should be enough to look at the smallest positive literal and the largest negative literal, since these are the hardest to eliminate. Therefore we define r(C)r(C) to be the following operation on a clause: literals over WW variables are left untouched, all positive XX literals but the smallest are removed, and all negative XX literals but the largest are removed. Formally,

r(iAxiiBxi¯(i,j)Cwi,jbi,j)=xminAxmaxB¯(i,j)Cwi,jbi,jr\biggl{(}\bigvee_{i\in A}x_{i}\lor\bigvee_{i\in B}\overline{x_{i}}\lor\bigvee_{(i,j)\in C}w_{i,j}^{b_{i,j}}\biggr{)}=x_{\min A}\lor\overline{x_{\max B}}\lor\bigvee_{(i,j)\in C}w_{i,j}^{b_{i,j}} (12)

where xminAx_{\min A} (resp. xmaxB¯\overline{x_{\max B}}) is omitted if AA (resp. BB) is empty.

We need the following property of r(C)r(C).

Lemma 5.7.

If C\upharpoonrightσi1{C}{\upharpoonright_{\sigma_{i}}}\neq 1 and vars(r(C))Xi=\mathrm{vars(r(C))}\cap X_{i}=\emptyset then C\upharpoonrightσi{C}{\upharpoonright_{\sigma_{i}}} is supported over WW variables.

Proof.

The hypothesis that vars(r(C))Xi=\mathrm{vars(r(C))}\cap X_{i}=\emptyset implies that the smallest positive XX literal in CC is either not larger than ii\ell or larger than (i+1)(i+1)\ell, but the hypothesis that C\upharpoonrightσi1{C}{\upharpoonright_{\sigma_{i}}}\neq 1 rules out the first case. Therefore all positive XX literals are falsified by σi\sigma_{i}. Analogously the largest negative XX literal is not larger than ii\ell and all negative XX literals are also falsified. ∎

We define each part ψi\psi_{i} to consist of all clauses CσC\in\sigma such that CC is

  1. 1.

    an 𝒳\mathcal{X} axiom not satisfied by σi\sigma_{i}; or

  2. 2.

    the conclusion of an inference with pivot in XiX_{i}; or

  3. 3.

    the conclusion of an inference with pivot in WW that depends on an 𝒳\mathcal{X} axiom if r(C)r(C) contains a variable in XiX_{i}; or

  4. 4.

    the conclusion of an inference with pivot in WW that does not depend on 𝒳\mathcal{X} axioms if the only immediate successor of CC is in ψi\psi_{i}.

This is the point in the proof where we use crucially that the original derivation is in RMA form: because clauses that do not depend on 𝒳\mathcal{X} axioms are not merges, they have only one successor and the definition is well-formed.

Ideally we would like to argue that parts ψi\psi_{i} are pairwise disjoint. This is not quite true, but nevertheless they do not overlap too much.

Lemma 5.8.

Let ψ\psi and {ψii[]}\{\psi_{i}\mid i\in[\ell]\} be as discussed above. Then 2|ψ|i|ψi|2\lvert\psi\rvert\geq\sum_{i}\lvert\psi_{i}\rvert.

Proof.

Axioms may appear in at most two different ψi\psi_{i}, and clauses obtained after resolving with an XX pivot in only one. The only other clauses that depend on an 𝒳\mathcal{X} axiom and may appear in different ψi\psi_{i} are obtained after resolving with a WW pivot, but since r(C)r(C) only contains two XX variables, such clause only may appear in two different ψi\psi_{i}. Finally, clauses that do not depend on an 𝒳\mathcal{X} axiom appear in the same ψi\psi_{i} as one clause of the previous types, and therefore at most two different parts. ∎

To conclude the proof we need to argue that each ψi\psi_{i} is large. The intuitive reason is that ψi\psi_{i} must use one 𝒳\mathcal{X} axiom for each j[(i,(i+1)]j\in[(i\ell,(i+1)\ell], which introduces a pair of WW variables from each WjW_{j} block, but since no clause contains more than /8\ell/8 such variables, we need to use enough 𝒲\mathcal{W} axioms to remove the aforementioned WW variables. Formally the claim follows from these two lemmas.

Lemma 5.9.

For each i[]i\in[\ell] there exists an integer tit_{i} such that s(ψi[1,ti]\upharpoonrightσi)s({\psi_{i}[1,t_{i}]}{\upharpoonright_{\sigma_{i}}}) is a resolution derivation of a clause supported on WW variables that depends on an 𝒳\mathcal{X} axiom.

Proof.

Let CtiC_{t_{i}} be the first clause in ψi\psi_{i} that depends on an 𝒳\mathcal{X} axiom and such that Cti\upharpoonrightσi{C_{t_{i}}}{\upharpoonright_{\sigma_{i}}} is supported on WW variables. We prove that tit_{i} is well-defined, that ψi[1,ti]\upharpoonrightσi{\psi_{i}[1,t_{i}]}{\upharpoonright_{\sigma_{i}}} is a valid semantic resolution derivation, and that Dti=s(ψi\upharpoonrightσi)D_{t_{i}}=s({\psi_{i}}{\upharpoonright_{\sigma_{i}}}) depends on an 𝒳\mathcal{X} axiom.

Our induction hypothesis is that for ktik\leq t_{i} (or any kk if tit_{i} does not exist), if the clause CkψC_{k}\in\psi depends on an 𝒳\mathcal{X} axiom and is not satisfied by σi\sigma_{i}, then there exists a clause CkψiC_{k^{\prime}}\in\psi_{i} with kkk^{\prime}\leq k that implies CkC_{k} modulo σi\sigma_{i}, that is Ck\upharpoonrightσi\vDashCk\upharpoonrightσi{C_{k^{\prime}}}{\upharpoonright_{\sigma_{i}}}\vDash{C_{k}}{\upharpoonright_{\sigma_{i}}}, and depends on an 𝒳\mathcal{X} axiom (over ψ\psi).

If the induction hypothesis holds then tit_{i} is well-defined: since CtC_{t} is not satisfied by σi\sigma_{i} and depends on an 𝒳\mathcal{X} axiom there exists a clause CtψiC_{t^{\prime}}\in\psi_{i} that depends on an 𝒳\mathcal{X} axiom and such that Ct\upharpoonrightσi\vDashCt\upharpoonrightσi=Ct{C_{t^{\prime}}}{\upharpoonright_{\sigma_{i}}}\vDash{C_{t}}{\upharpoonright_{\sigma_{i}}}=C_{t}, which is supported on WW variables.

The base case is when CkC_{k} is a non-satisfied 𝒳\mathcal{X} axiom, where we can take Ck=CkC_{k^{\prime}}=C_{k}. For the inductive case let CpC_{p} and CqC_{q} be the premises of CkC_{k} in ψ\psi. If exactly one of the premises, say CpC_{p}, is non-satisfied and, furthermore, depends on an 𝒳\mathcal{X} axiom, then by the induction hypothesis we can take Ck=CpC_{k^{\prime}}=C_{p^{\prime}}. Otherwise we need to consider a few subcases. If the pivot is an XX variable then both premises depend on an 𝒳\mathcal{X} axiom (by Lemma 5.3), hence neither premise is satisfied. It follows that the pivot is unassigned by σi\sigma_{i}, and therefore we can take Ck=CkC_{k^{\prime}}=C_{k}.

If the pivot is a WW variable then, because σi\sigma_{i} only assigns XX variables, neither premise is satisfied. We have two subcases: if exactly one premise depends on an 𝒳\mathcal{X} axiom, say CpC_{p}, then CpC_{p^{\prime}} is present in ψi\psi_{i}, and by construction of ψi\psi_{i} the other premise CqC_{q} is present in ψi\psi_{i} if and only if the conclusion CkC_{k} is. If both premises depend on an 𝒳\mathcal{X} axiom then both CpC_{p^{\prime}} and CqC_{q^{\prime}} are present in ψi\psi_{i}.

Therefore in the two latter subcases it is enough to prove that CkψiC_{k}\in\psi_{i}, since then we can take Ck=CkC_{k^{\prime}}=C_{k} and we have that Ck\upharpoonrightσi{C_{k}}{\upharpoonright_{\sigma_{i}}} follows from a valid semantic resolution step. Indeed by Lemma 5.7 Ck\upharpoonrightσi{C_{k}}{\upharpoonright_{\sigma_{i}}} is a clause supported on WW variables, which by definition of CtiC_{t_{i}} implies that k=tik=t_{i}. However, since the pivot is a WW variable, Cp\upharpoonrightσi{C_{p^{\prime}}}{\upharpoonright_{\sigma_{i}}} is also supported on WW variables and, together with the fact that CpC_{p^{\prime}} depends on an 𝒳\mathcal{X} axiom, this contradicts that CtiC_{t_{i}} is the first such clause.

This finishes the first induction argument and proves that ψ[1,ti]\upharpoonrightσi{\psi[1,t_{i}]}{\upharpoonright_{\sigma_{i}}} is a valid semantic derivation; it remains to prove that DtiD_{t_{i}} depends on an XX axiom over s(ψi\upharpoonrightσi)s({\psi_{i}}{\upharpoonright_{\sigma_{i}}}). We prove by a second induction argument that for every clause Dks(ψi[1,ti]\upharpoonrightσi)D_{k}\in s({\psi_{i}[1,t_{i}]}{\upharpoonright_{\sigma_{i}}}), if CkC_{k} depends on an 𝒳\mathcal{X} axiom then so does DkD_{k}. The base case, when DkD_{k} is an axiom, holds.

For the inductive case fix CkC_{k}, Ek=Ck\upharpoonrightσiE_{k}={C_{k}}{\upharpoonright_{\sigma_{i}}}, and Dk=s(Ek)D_{k}=s(E_{k}), and let Ep=Cp\upharpoonrightσiE_{p}={C_{p}}{\upharpoonright_{\sigma_{i}}} and Eq=Cq\upharpoonrightσiE_{q}={C_{q}}{\upharpoonright_{\sigma_{i}}} be the premises of EkE_{k} in ψi\upharpoonrightσ{\psi_{i}}{\upharpoonright_{\sigma}}. When both CpC_{p} and CqC_{q} depend on an XX axiom, then by hypothesis so do DpD_{p} and DqD_{q} and we are done. We only need to argue the case when one premise CpC_{p} depends on an XX axiom and the other premise CqC_{q} does not. In that case, because σi\sigma_{i} only affects XX variables, all the axioms used in the derivation of CqC_{q} are left untouched by σi\sigma_{i}, therefore we have that s(σi(Cq))=Cqs(\sigma_{i}(C_{q}))=C_{q}, which contains the pivot used to derive CkC_{k} and therefore does not imply s(σi(Ck))s(\sigma_{i}(C_{k})). By construction of s()s(\cdot), s(σi(Ck))s(\sigma_{i}(C_{k})) depends on s(σi(Cp))s(\sigma_{i}(C_{p})). ∎

Lemma 5.10.

Let η\eta be a resolution derivation from F,1,nF_{\ell,1,n} of a clause CC supported on WW variables that depends on an 𝒳\mathcal{X} axiom. Then |η|(n2)(μ(C))/2\lvert\eta\rvert\geq(n-2)(\ell-\mu(C))/2.

Proof.

By Lemma 5.4 we can assume that η\eta uses at least one Aj,bA_{j,b} axiom for each j[]j\in[\ell].

Let J={j[]wj,kvars(C)}J=\{j\in[\ell]\mid\exists w_{j,k}\in\mathrm{vars(C)}\} be the set of WW blocks mentioned by CC. We show that for each jJ¯=[]Jj\in\overline{J}=[\ell]\setminus J at least (n2)/2(n-2)/2 axioms over variables in WjW_{j} appear in η\eta, which makes for at least (n2)|J¯|/2=(n2)(μ(C))/2(n-2)\lvert\overline{J}\rvert/2=(n-2)(\ell-\mu(C))/2 axioms.

Fix jJ¯j\in\overline{J} and assume for the sake of contradiction that less than (n2)/2(n-2)/2 axioms over variables in WjW_{j} appear in η\eta. Then there exists k[2,n1]k\in[2,n-1] such that variable wj,kw_{j,k} does not appear in η\eta. Rename variables as follows: wj,kykw_{j,k^{\prime}}\mapsto y_{k^{\prime}} for k<kk^{\prime}<k, and wj,kykn¯w_{j,k^{\prime}}\mapsto\overline{y_{k^{\prime}-n}} for k>kk^{\prime}>k. Then we can prove by induction, analogously to the proof of Lemma 5.3, that every clause derived from axiom Aj,bA_{j,b} is of the form ykyk′′¯Dy_{k^{\prime}}\lor\overline{y_{k^{\prime\prime}}}\lor D where DD are literals supported outside WjW_{j}. Since that includes CC, it contradicts our assumption that jJj\notin J. ∎

To conclude the proof of Theorem 5.1 we simply need to put the pieces together.

Proof of Theorem 5.1.

We take as the formula family F=48logn,n,nF_{\ell=48\log n,n,n}, for which a resolution refutation of length O(nlogn)\mathrm{O}(n\log n) exists by Lemma 5.2.

To prove a lower bound we and assume that a RMA refutation π\pi of length Ln3=216=o((4/3)8)L\leq n^{3}=2^{16\ell}=\mathrm{o}((4/3)^{8\ell}) exists; otherwise the lower bound trivially holds. We apply the restriction given by Lemma 5.5 to π\pi and we use Lemma 5.6 to obtain a resolution derivation ψ\psi of a clause supported on WW variables that uses an 𝒳\mathcal{X} axiom. We then break ψ\psi into mm parts ψi\psi_{i}, each of size at least n/16n\ell/16 as follows from Lemmas 5.9 and 5.10. Finally by Lemma 5.8 we have |π||ψ|mn/32=Ω(n2logn)\lvert\pi\rvert\geq\lvert\psi\rvert\geq mn\ell/32=\Omega(n^{2}\log n). ∎

5.4 Structural Consequences

Theorem 5.1 immediately gives us two structural properties of RML and RMA. One is that proof length may decrease when introducing a weakening rule.

Corollary 5.11.

There exists a family of formulas over O(nlogn)\mathrm{O}(n\log n) variables and O(nlogn)\mathrm{O}(n\log n) clauses that have RML with weakening refutations of length O(nlogn)\mathrm{O}(n\log n) but every RMA refutation requires length Ω(n2logn)\Omega(n^{2}\log n).

Proof.

Consider the formula Fnz¯F_{n}\land\overline{z}, where FnF_{n} is the formula given by Theorem 5.1 and zz is a new variable. If we weaken every clause CFnC\in F_{n} to CzC\lor z then we can derive FzzF\lor z\vdash z in O(nlogn)\mathrm{O}(n\log n) RML steps because each inference is a merge. However, if we cannot do weakening, then z¯\overline{z} cannot be resolved with any clause in FnF_{n} and the lower bound of Theorem 5.1 applies. ∎

The second property is that RML and RMA are not natural proof systems in the sense of [BKS04] because proof length may increase after a restriction.

Corollary 5.12.

There exists a restriction ρ\rho and a family of formulas over O(nlogn)\mathrm{O}(n\log n) variables and O(nlogn)\mathrm{O}(n\log n) clauses that have RML refutations of length O(nlogn)\mathrm{O}(n\log n) but every RMA refutation of Fn\upharpoonrightρ{F_{n}}{\upharpoonright_{\rho}} requires length Ω(n2logn)\Omega(n^{2}\log n).

Proof.

Consider the formula Gn=(Fnz)z¯G_{n}=(F_{n}\lor z)\land\overline{z}, where FnF_{n} is the formula given by Theorem 5.1, Fz={CzCF}F\lor z=\{C\lor z\mid C\in F\}, and zz is a new variable. As in the proof of Corollary 5.11 there is a RML derivation of zz of length O(nlogn)\mathrm{O}(n\log n) steps, while Gn\upharpoonrightρ=Fn{G_{n}}{\upharpoonright_{\rho}}=F_{n}. ∎

6 Further Separations

We can separate the different flavours of merge resolution that we introduced using a few variations of F,m,nF_{\ell,m,n} where we add a constant number of redundant clauses for each i[]i\in[\ell]. We consider these different clauses part of 𝒲\mathcal{W}.

Upper bounds all follow the same pattern. We first show on a case-by-case basis how to obtain w1wn¯w_{1}\overline{w_{n}} and w1¯wn\overline{w_{1}}w_{n} as lemmas, and then proceed as in Section 5.2.

Towards proving lower bounds we are going to generalize the lower bound part of the proof of Theorem 5.1 to apply to these variations as well. Fortunately we only require a few local modifications.

First, we need to prove an equivalent of Lemma 5.3, which we do on a case-by-case basis.

Second, we need to show that kk-respecting restrictions can be extended to the new variables. For each block JiJ_{i}, since the new clauses are semantically subsumed by wi,1=wi,2w_{i,1}=w_{i,2}, there exists a way to map the new variables into wi,1w_{i,1} and wi,2w_{i,2} so that the result of the restriction is the same as if we had started with clauses wi,1¯wi,2\overline{w_{i,1}}\lor w_{i,2} and wi,1wi,2¯w_{i,1}\lor\overline{w_{i,2}}, which are already part of 𝒲i\mathcal{W}_{i}. That is, the formula that we work with after Lemma 5.6 is a copy of an unaltered F,m,nF_{\ell^{\prime},m^{\prime},n^{\prime}} formula.

The only part of the lower bound that depends on the specific subsystem of Resolution is Lemma 5.6; afterwards all the information we use is that no clause supported on WW variables is reused. Furthermore, the only property of the subsystem that we use in the proof of Lemma 5.6 is that Lemma 5.3 applies. Therefore, the modifications we just outlined are sufficient for the lower bound to go through.

6.1 Separation between RMA and LRMA

Proposition 6.1.

There exists a family of formulas over O(nlogn)\mathrm{O}(n\log n) variables and O(nlogn)\mathrm{O}(n\log n) clauses that have RMA refutations of length O(nlogn)\mathrm{O}(n\log n) but every LRMA refutation requires length Ω(n2logn)\Omega(n^{2}\log n).

The separating formula is Fm,n,(1)F_{m,n,\ell}^{(1)}, where we add to Fm,n,F_{m,n,\ell} clauses

wi,1¯wi,2zi¯,\displaystyle\overline{w_{i,1}}\lor w_{i,2}\lor\overline{z_{i}}, (C1)
wi,2¯zi,\displaystyle\overline{w_{i,2}}\lor z_{i}, (C2)
wi,1wi,2¯yi¯,\displaystyle w_{i,1}\lor\overline{w_{i,2}}\lor\overline{y_{i}}, (C3)
wi,2yi,\displaystyle w_{i,2}\lor y_{i}, (C4)

for each i[]i\in[\ell]. The new variables can be assigned as zi=wi,1z_{i}=w_{i,1} and yi=wi,1¯y_{i}=\overline{w_{i,1}} to obtain the original formula back.

The upper bound follows from the following lemma.

Lemma 6.2.

Clauses wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} and wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n} can be derived as lemmas from Fm,n,(1)F_{m,n,\ell}^{(1)} in length O(n)\mathrm{O}(n) in RMA.

Proof.

We resolve clause wi,1¯wi,2\overline{w_{i,1}}\lor w_{i,2} first with (C2) and then (C1) in order to obtain wi,1¯wi,2\overline{w_{i,1}}\lor w_{i,2} as a merge, then derive wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n}, having a merge as its ancestor, so it can be remembered. Analogously starting from wi,1wi,2¯w_{i,1}\lor\overline{w_{i,2}}, (C3), and (C4) we can obtain wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} as a lemma. ∎

The following observation is useful for the lower bound.

Lemma 6.3.

Let CC and DD be clauses with two pairs of opposite literals. Then CC and DD cannot appear in the same locally regular input derivation.

Proof.

Let C=xyCC=x\lor y\lor C^{\prime} and D=x¯y¯DD=\overline{x}\lor\overline{y}\lor D^{\prime}. Assume wlog that CC is the first clause out of CC and DD to appear in the derivation. If xx or yy are used as pivots before DD, then the locally regular condition prevents using DD as an axiom. Otherwise xyx\lor y appears in the derivation since the time CC is used, which also prevents using DD. ∎

The equivalent of Lemma 5.3 is the following.

Lemma 6.4.

Let η\eta be a LRMA derivation that only depends on 𝒲\mathcal{W} axioms. Then no clause in η\eta can be reused.

Proof.

We can only obtain a merge using one of (C1) or (C3), assume wlog (C1) is the first of these to be used in the derivation. By Lemma 6.3 neither (C2) nor (C3) appear in the derivation. We can show by induction that we can only obtain clauses of the form wi,j¯wi,j¯zi¯\overline{w_{i,j}}\lor\overline{w_{i,j^{\prime}}}\lor\overline{z_{i}} or yiwi,j¯zi¯y_{i}\lor\overline{w_{i,j}}\lor\overline{z_{i}}, never as a merge. ∎

6.2 Separation between RML/LRMA and LRML

Proposition 6.5.

There exists a family of formulas over O(nlogn)\mathrm{O}(n\log n) variables and O(nlogn)\mathrm{O}(n\log n) clauses that have RML and LRMA and refutations of length O(nlogn)\mathrm{O}(n\log n) but every LRML refutation requires length Ω(n2logn)\Omega(n^{2}\log n).

The separating formula is Fm,n,(2)F_{m,n,\ell}^{(2)}, where we add to Fm,n,F_{m,n,\ell} clauses

ziwi,1¯wi,2,\displaystyle z_{i}\lor\overline{w_{i,1}}\lor w_{i,2}, (C1)
zi¯wi,1¯wi,2,\displaystyle\overline{z_{i}}\lor\overline{w_{i,1}}\lor w_{i,2}, (C2)
yiwi,1wi,2¯,\displaystyle y_{i}\lor w_{i,1}\lor\overline{w_{i,2}}, (C3)
yi¯wi,1wi,2¯,\displaystyle\overline{y_{i}}\lor w_{i,1}\lor\overline{w_{i,2}}, (C4)

for each i[]i\in[\ell]. The new variables can be assigned as zi=1z_{i}=1 and yi=1y_{i}=1 to obtain the original formula back.

The upper bounds follow respectively from the following lemmas.

Lemma 6.6.

Clauses wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} and wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n} can be derived as lemmas from Fm,n,(2)F_{m,n,\ell}^{(2)} in length O(n)\mathrm{O}(n) in RML.

Proof.

We first resolve clauses wn1¯wi,n\overline{w_{n-1}}\lor w_{i,n}, wi,n2¯wi,n1\overline{w_{i,n-2}}\lor w_{i,n-1}, …, wi,2¯wi,3\overline{w_{i,2}}\lor w_{i,3}, (C1) to obtain ziwi,1¯wi,nz_{i}\lor\overline{w_{i,1}}\lor w_{i,n}. We continue the input derivation resolving with (C2) to obtain wi,1¯wi,2wi,n\overline{w_{i,1}}\lor w_{i,2}\lor w_{i,n}. We then resolve with wi,2¯wi,3\overline{w_{i,2}}\lor w_{i,3}, wi,3¯wi,4\overline{w_{i,3}}\lor w_{i,4}, …, wi,n1¯wi,n\overline{w_{i,n-1}}\lor w_{i,n} to obtain wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n} as a merge over wi,nw_{i,n}. Analogously we can obtain wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}}. ∎

Lemma 6.7.

Clauses wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} and wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n} can be derived as lemmas from Fm,n,(2)F_{m,n,\ell}^{(2)} in length O(n)\mathrm{O}(n) in LRMA.

Proof.

We resolve clauses (C1) and (C2) to obtain wi,1¯wi,2\overline{w_{i,1}}\lor w_{i,2}, which is a merge, then derive wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n}, having a merge as its ancestor, so it can be used as a lemma. Analogously starting from (C3) and (C4) we can obtain wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} as a lemma. ∎

The equivalent of Lemma 5.3 is the following.

Lemma 6.8.

Let η\eta be a LRML derivation that only depends on 𝒲\mathcal{W} axioms. Then no clause in η\eta can be reused.

The proof idea is that the only merge we can obtain involves the ziz_{i} or the yiy_{i} variable. If we just resolve the two clauses over such a variable we obtain a clause we already had, so this is useless. Otherwise we are resolving one of w2w_{2} away, which would be reintroduced at the time of resolving ziz_{i} away, and that is not allowed by the SR condition.

Proof.

We can only obtain a merge by using one of the new clauses (C1)–(C4). If we resolve either pair of clauses over yiy_{i} or over ziz_{i} then we obtain a clause that was already present in the formula, and therefore we may preprocess such derivation away.

Otherwise consider the first step in the derivation where one of the new clauses is used as a premise, assume wlog it is (C1). That step is with a clause of the form wi,2¯wi,j\overline{w_{i,2}}\lor w_{i,j}, and we obtain a clause of the form ziwi,1¯wi,jz_{i}\lor\overline{w_{i,1}}\lor w_{i,j}, which is not a merge. That clause can be possibly resolved over wi,jw_{i,j} (j>2j>2) to obtain other clauses of the same form, neither of which is a merge, but it cannot be resolved over yiy_{i}, ziz_{i}, or wi,1w_{i,1} because that step would reintroduce variable wi,2w_{i,2}. ∎

6.3 Separation between LRML and REL

Proposition 6.9.

There exists a family of formulas over O(nlogn)\mathrm{O}(n\log n) variables and O(nlogn)\mathrm{O}(n\log n) clauses that have LRML refutations of length O(nlogn)\mathrm{O}(n\log n) but every REL refutation requires length Ω(n2logn)\Omega(n^{2}\log n).

The separating formula is Fm,n,(3)F_{m,n,\ell}^{(3)}, where we add to Fm,n,F_{m,n,\ell} clauses

wi,1¯wi,2¯wi,3,\displaystyle\overline{w_{i,1}}\lor\overline{w_{i,2}}\lor w_{i,3}, (C1)
wi,1wi,2wi,3¯\displaystyle w_{i,1}\lor w_{i,2}\lor\overline{w_{i,3}} (C2)

for each i[]i\in[\ell]. If we assign wi,2=wi,1w_{i,2}=w_{i,1} we obtain a copy of Fm,n1,F_{m,n-1,\ell} which, even if technically it is not the same formula we started with, is enough for our purposes.

The upper bound follows from the following lemma.

Lemma 6.10.

Clauses wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} and wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n} can be derived as lemmas from Fm,n,(3)F_{m,n,\ell}^{(3)} in length O(n)\mathrm{O}(n) in LRML.

Proof.

We resolve (C1) with wi,3¯wi,4\overline{w_{i,3}}\lor w_{i,4}, …, wi,n1¯wi,n\overline{w_{i,n-1}}\lor w_{i,n} to obtain wi,1¯wi,2¯wi,n\overline{w_{i,1}}\lor\overline{w_{i,2}}\lor w_{i,n}, then with wi,1¯wi,2\overline{w_{i,1}}\lor w_{i,2} to obtain wi,1¯wi,n\overline{w_{i,1}}\lor w_{i,n} as a merge. Analogously starting from (C2) we can obtain wi,1wi,n¯w_{i,1}\lor\overline{w_{i,n}} as a lemma. ∎

The equivalent of Lemma 5.3 is the following.

Lemma 6.11.

Let η\eta be a REL derivation that only depends on 𝒲\mathcal{W} axioms. Then no clause in η\eta can be reused.

Proof.

Observe that every derivable clause has width at least 22. Let CC be any derivable clause and \ell any literal in CC. We have that α=C¯\alpha=\overline{C\setminus\ell} is not empty. However, assigning any variable wi,jw_{i,j} immediately propagates all variables, hence \ell is not empowering. ∎

7 Concluding Remarks

In this paper, we address the question of the tightness of simulation of resolution proofs by CDCL solvers. Specifically, we show that RMA, among other flavours of DAG-like merge resolution, simulates standard resolution with at most a linear multiplicative overhead. However, contrary to what we see in the tree-like case, this overhead is necessary. While the proof systems we introduce help us explain one source of overhead in the simulation of resolution by CDCL, it is not clear if they capture it exactly. In other words, an interesting future direction would be to explore whether it is possible for CDCL to simulate some flavour of merge resolution with less overhead than what is required to simulate standard resolution.

Acknowledgements

The authors are grateful to Yuval Filmus and a long list of participants in the program Satisfiability: Theory, Practice, and Beyond at the Simons Institute for the Theory of Computing for numerous discussions. This work was done in part while the authors were visiting the Simons Institute for the Theory of Computing.

Appendix A Tree-like Merge Resolution

For completeness we informally sketch the proofs of Lemma 3.2 and Theorem 3.3, which can be found in full detail in [And68].

Lemma A.1 (Lemma 3.2, restated).

If there is a tree-like resolution derivation of CC of length LL where at most the root is a merge, then there is an input resolution derivation of some CCC^{\prime}\subseteq C of length at most LL.

Proof (sketch).

We prove by induction on |η|\lvert\eta\rvert that for every axiom EηE\in\eta there exists an input derivation of CC^{\prime} that uses a subset of the axioms of η\eta where EE is the topmost axiom. As intermediate objects we allow clauses in this derivation to contain opposite literals; these are cleaned up later.

Let C=Res(Ax,Bx¯)C=\mathrm{Res}(A\lor x,B\lor\overline{x}), and let η1\eta_{1} and η2\eta_{2} be the derivations used to infer AxA\lor x and BxB\lor x respectively. Assume wlog that Eη1E\in\eta_{1}. Since η2\eta_{2} does not contain any merges there exists a unique path from Bx¯B\lor\overline{x} to an axiom Dx¯η2D\lor\overline{x}\in\eta_{2}, where all clauses contain x¯\overline{x}. Note that other clauses in η2\eta_{2} might still contain xx or x¯\overline{x}. We replace Dx¯D\lor\overline{x} by DD in η2\eta_{2} (and consequently remove all the occurrences of x¯\overline{x} in the aforementioned path) and we obtain a valid derivation η3\eta_{3} of BB. We apply the induction hypothesis to η1\eta_{1} and η3\eta_{3} to obtain two unit derivations η4\eta_{4} and η5\eta_{5} of AxAxA^{\prime}\lor x\subseteq A\lor x and BBB^{\prime}\subseteq B whose topmost leaves are EE and DD. We replace DD by DAD\lor A^{\prime} in η5\eta_{5} and obtain a unit derivation η6\eta_{6} of BA′′BAB^{\prime}\lor A^{\prime\prime}\subseteq B^{\prime}\lor A^{\prime}. We stitch together η4\eta_{4} and η6\eta_{6} by observing that Res(Ax,Dx¯)=AD\mathrm{Res}(A^{\prime}\lor x,D\lor\overline{x})=A^{\prime}\lor D, which is the only axiom in η6\eta_{6} not present in the original axioms, and obtain a unit derivation η7\eta_{7} of BA′′=CCB\lor A^{\prime\prime}=C^{\prime}\subseteq C that only uses original axioms.

Finally, and outside the inductive argument, we get rid of clauses that contain opposite literals by replacing any such clause by 11 to obtain a semantic derivation η8\eta_{8}. Its syntactic counterpart s(η8)s(\eta_{8}) satisfies the conclusion of the lemma. ∎

Refer to caption
Figure 2: Proof of Lemma A.1
Theorem A.2 (Theorem 3.3, restated).

If there is a tree-like resolution derivation of CC of length LL, then there is a merge resolution derivation of some CCC^{\prime}\subseteq C of length at most LL.

Proof (sketch).

The proof is by induction on the number of merges. The base case when there are no merges follows by Lemma A.1. Otherwise let ψ\psi be a subtree where exactly the root CC is a merge. Let ψ\psi^{\prime} be the input resolution derivation of CC^{\prime} given by Lemma A.1, let DD be the last merge in ψ\psi^{\prime}, and let ω\omega and ω\omega^{\prime} be the fragments of ψ\psi^{\prime} from DD to CC^{\prime} and up to DD respectively. We replace ψ\psi by ω\omega in η\eta to obtain a refutation η\eta^{\prime} that uses DD as an axiom (note that in replacing CC by CC^{\prime} we may have to prune away parts of η\eta). Because η\eta^{\prime} has one less merge we can apply the induction hypothesis and obtain a merge resolution derivation ψ′′\psi^{\prime\prime}. Finally we replace the axiom DD by the derivation ω\omega^{\prime}. ∎

Refer to caption
Figure 3: Proof of Theorem A.2

References

  • [\lfstyleABH+08] Gilles Audemard, Lucas Bordeaux, Youssef Hamadi, Saïd Jabbour, and Lakhdar Sais. A generalized framework for conflict analysis. In Hans Kleine Büning and Xishun Zhao, editors, Theory and Applications of Satisfiability Testing - SAT 2008, 11th International Conference, SAT 2008, Guangzhou, China, May 12-15, 2008. Proceedings, volume 4996 of Lecture Notes in Computer Science, pages 21–27. Springer, 2008. doi:10.1007/978-3-540-79719-7\_3.
  • [\lfstyleAFT11] Albert Atserias, Johannes Klaus Fichte, and Marc Thurley. Clause-learning algorithms with many restarts and bounded-width resolution. Journal of Artificial Intelligence Research, 40:353–373, January 2011. Preliminary version in SAT ’09.
  • [\lfstyleAnd68] Peter B. Andrews. Resolution with merging. J. ACM, 15(3):367–381, 1968.
  • [\lfstyleBB21] Olaf Beyersdorff and Benjamin Böhm. Understanding the relative strength of QBF CDCL solvers and QBF resolution. In James R. Lee, editor, 12th Innovations in Theoretical Computer Science Conference, ITCS 2021, January 6-8, 2021, Virtual Conference, volume 185 of LIPIcs, pages 12:1–12:20. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2021. doi:10.4230/LIPIcs.ITCS.2021.12.
  • [\lfstyleBF97] Avrim L Blum and Merrick L Furst. Fast planning through planning graph analysis. Artificial intelligence, 90(1-2):281–300, 1997.
  • [\lfstyleBHJ08] Samuel R. Buss, Jan Hoffmann, and Jan Johannsen. Resolution trees with lemmas: Resolution refinements that characterize DLL-algorithms with clause learning. Logical Methods in Computer Science, 4(4:13), December 2008.
  • [\lfstyleBKS04] Paul Beame, Henry Kautz, and Ashish Sabharwal. Towards understanding and harnessing the potential of clause learning. Journal of Artificial Intelligence Research, 22:319–351, December 2004. Preliminary version in IJCAI ’03.
  • [\lfstyleBN21] Sam Buss and Jakob Nordström. Proof complexity and SAT solving. In Armin Biere, Marijn Heule, Hans van Maaren, and Toby Walsh, editors, Handbook of Satisfiability, volume 336 of Frontiers in Artificial Intelligence and Applications, chapter 7, pages 233–350. IOS Press, 2nd edition, 2021. doi:10.3233/FAIA200990.
  • [\lfstyleBS97] Roberto J. Bayardo Jr. and Robert Schrag. Using CSP look-back techniques to solve real-world SAT instances. In Proceedings of the 14th National Conference on Artificial Intelligence (AAAI ’97), pages 203–208, July 1997.
  • [\lfstyleCGP+08] Cristian Cadar, Vijay Ganesh, Peter M Pawlowski, David L Dill, and Dawson R Engler. EXE: Automatically Generating Inputs of Death. ACM Transactions on Information and System Security (TISSEC), 12(2):1–38, 2008.
  • [\lfstyleDHN07] Nachum Dershowitz, Ziyad Hanna, and Alexander Nadel. Towards a better understanding of the functionality of a conflict-driven SAT solver. In João Marques-Silva and Karem A. Sakallah, editors, Theory and Applications of Satisfiability Testing - SAT 2007, 10th International Conference, Lisbon, Portugal, May 28-31, 2007, Proceedings, volume 4501 of Lecture Notes in Computer Science, pages 287–293. Springer, 2007. doi:10.1007/978-3-540-72788-0\_27.
  • [\lfstyledV94] Alvaro del Val. Tractable databases: How to make propositional unit resolution complete through compilation. In Jon Doyle, Erik Sandewall, and Pietro Torasso, editors, Proceedings of the 4th International Conference on Principles of Knowledge Representation and Reasoning (KR’94). Bonn, Germany, May 24-27, 1994, pages 551–561. Morgan Kaufmann, 1994.
  • [\lfstyleDVT07] Julian Dolby, Mandana Vaziri, and Frank Tip. Finding Bugs Efficiently With a SAT Solver. In Proceedings of the 6th joint meeting of the European Software Engineering Conference and the ACM SIGSOFT International Symposium on Foundations of Software Engineering, pages 195–204, 2007. doi:10.1145/1287624.1287653.
  • [\lfstyleFB20] Nick Feng and Fahiem Bacchus. Clause size reduction with all-uip learning. In Luca Pulina and Martina Seidl, editors, Theory and Applications of Satisfiability Testing - SAT 2020 - 23rd International Conference, Alghero, Italy, July 3-10, 2020, Proceedings, volume 12178 of Lecture Notes in Computer Science, pages 28–45. Springer, 2020. doi:10.1007/978-3-030-51825-7\_3.
  • [\lfstyleHBPV08] Philipp Hertel, Fahiem Bacchus, Toniann Pitassi, and Allen Van Gelder. Clause learning can effectively P-simulate general propositional resolution. In Proceedings of the 23rd National Conference on Artificial Intelligence (AAAI ’08), pages 283–290, July 2008.
  • [\lfstyleLFV+20] Chunxiao Li, Noah Fleming, Marc Vinyals, Toniann Pitassi, and Vijay Ganesh. Towards a complexity-theoretic understanding of restarts in sat solvers. In Theory and Applications of Satisfiability Testing–SAT 2020: 23rd International Conference, Alghero, Italy, July 3–10, 2020, Proceedings 23, pages 233–249. Springer, 2020.
  • [\lfstyleMLM21] João Marques-Silva, Inês Lynce, and Sharad Malik. Conflict-driven clause learning SAT solvers. In Armin Biere, Marijn Heule, Hans van Maaren, and Toby Walsh, editors, Handbook of Satisfiability - Second Edition, volume 336 of Frontiers in Artificial Intelligence and Applications, pages 133–182. IOS Press, 2021. doi:10.3233/FAIA200987.
  • [\lfstyleMMZ+01] Matthew W. Moskewicz, Conor F. Madigan, Ying Zhao, Lintao Zhang, and Sharad Malik. Chaff: Engineering an efficient SAT solver. In Proceedings of the 38th Design Automation Conference (DAC ’01), pages 530–535, June 2001.
  • [\lfstyleMPR20] Nathan Mull, Shuo Pang, and Alexander A. Razborov. On CDCL-based proof systems with the ordered decision strategy. In Proceedings of the 23rd International Conference on Theory and Applications of Satisfiability Testing (SAT ’20), volume 12178 of Lecture Notes in Computer Science, pages 149–165. Springer, July 2020.
  • [\lfstyleMS85] Burkhard Monien and Ewald Speckenmeyer. Solving satisfiability in less than 2n2^{n} steps. Discret. Appl. Math., 10(3):287–295, 1985. doi:10.1016/0166-218X(85)90050-2.
  • [\lfstyleMS99] João P. Marques-Silva and Karem A. Sakallah. GRASP: A search algorithm for propositional satisfiability. IEEE Transactions on Computers, 48(5):506–521, May 1999. Preliminary version in ICCAD ’96.
  • [\lfstylePD08] Knot Pipatsrisawat and Adnan Darwiche. A new clause learning scheme for efficient unsatisfiability proofs. In Dieter Fox and Carla P. Gomes, editors, Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence, AAAI 2008, Chicago, Illinois, USA, July 13-17, 2008, pages 1481–1484. AAAI Press, 2008. URL: http://www.aaai.org/Library/AAAI/2008/aaai08-243.php.
  • [\lfstylePD11] Knot Pipatsrisawat and Adnan Darwiche. On the power of clause-learning SAT solvers as resolution engines. Artificial Intelligence, 175(2):512–525, February 2011. Preliminary version in CP ’09.
  • [\lfstyleRya04] Lawrence Ryan. Efficient algorithms for clause-learning SAT solvers. Master’s thesis, Simon Fraser University, 2004.
  • [\lfstyleSB09] Niklas Sörensson and Armin Biere. Minimizing learned clauses. In Proceedings of the 12th International Conference on Theory and Applications of Satisfiability Testing (SAT ’09), volume 5584 of Lecture Notes in Computer Science, pages 237–243. Springer, July 2009.
  • [\lfstyleVan05] Allen Van Gelder. Pool resolution and its relation to regular resolution and DPLL with clause learning. In Proceedings of the 12th International Conference on Logic for Programming, Artificial Intelligence, and Reasoning (LPAR ’05), volume 3835 of Lecture Notes in Computer Science, pages 580–594. Springer, 2005.
  • [\lfstyleVin20] Marc Vinyals. Hard examples for common variable decision heuristics. In Proceedings of the 34th AAAI Conference on Artificial Intelligence (AAAI ’20), pages 1652–1659, February 2020.
  • [\lfstyleXA05] Yichen Xie and Alexander Aiken. Saturn: A SAT-Based Tool for Bug Detection. In Proceedings of the 17th International Conference on Computer Aided Verification, CAV 2005, pages 139–143, 2005. doi:10.1007/11513988\_13.
  • [\lfstyleZMMM01] Lintao Zhang, Conor F. Madigan, Matthew W. Moskewicz, and Sharad Malik. Efficient conflict driven learning in Boolean satisfiability solver. In Proceedings of the IEEE/ACM International Conference on Computer-Aided Design (ICCAD ’01), pages 279–285, November 2001.
  • [\lfstyleZMW+18] Edward Zulkoski, Ruben Martins, Christoph M. Wintersteiger, Jia Hui Liang, Krzysztof Czarnecki, and Vijay Ganesh. The effect of structural measures and merges on SAT solver performance. In John N. Hooker, editor, Principles and Practice of Constraint Programming - 24th International Conference, CP 2018, Lille, France, August 27-31, 2018, Proceedings, volume 11008 of Lecture Notes in Computer Science, pages 436–452. Springer, 2018. doi:10.1007/978-3-319-98334-9\_29.