This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Department of Applied Mathematics, Faculty of Mathematics and Physics,
Charles University in Prague, Czech Republic
11email: kolman@kam.mff.cuni.cz, koutecky@kam.mff.cuni.cz

Extended Formulation for CSP that is Compact
for Instances of Bounded Treewidththanks: This research was partially supported by the project 14-10003S of GA ČR.

Petr Kolman    Martin Koutecký
Abstract

In this paper we provide an extended formulation for the class of constraint satisfaction problems and prove that its size is polynomial for instances whose constraint graph has bounded treewidth. This implies new upper bounds on extension complexity of several important NP-hard problems on graphs of bounded treewidth.

1 Introduction

Many important combinatorial optimization problems belong to the class of constraint satisfaction problems (CSP). Naturally, a lot of effort has been given to design efficient approximation algorithms for CSP, to prove complexity lower bounds for CSP, and to identify tractable instances of CSP (e.g., from the point of view of parameterized complexity). It has been shown that CSP is solvable in polynomial time for instances whose constraint graph has bounded treewidth [7].

In recent years, a lot of attention has been given to study extension complexity of problems [5]: what is the minimum number of inequalities representing a polytope whose (suitably chosen) linear projection coincides with the convex hull HH of all integral solutions of QQ? Such a polytope is called the extended formulation of HH. Note that membership of a problem in the class P of polynomially solvable problems does not necessarily imply the existence of an extended formulation of polynomial size [16]. In this work, we present an extended formulation for CSP and show that its size is polynomial for instances of CSP whose constraint graph has bounded treewidth.

1.1 Notation and Terminology

An instance Q=(V,𝒟,,𝒞)Q=(V,\mathcal{D},\mathcal{H},\mathcal{C}) of CSP consists of

  • a set of variables zvz_{v}, one for each vVv\in V; without loss of generality we assume that V={1,,n}V=\{1,\ldots,n\},

  • a set 𝒟\mathcal{D} of finite domains DvD_{v}\subseteq\mathbb{R} (also denoted D(v)D(v)), one for each vVv\in V,

  • a set of hard constraints {CU|UV}\mathcal{H}\subseteq\{C_{U}\ |\ U\subseteq V\} where each hard constraint CUC_{U}\in\mathcal{H} with U={i1,i2,,ik}U=\{i_{1},i_{2},\dots,i_{k}\} and i1<i2<<iki_{1}<i_{2}<\cdots<i_{k}, is a |U||U|-ary relation CUDi1×Di2××DikC_{U}\subseteq D_{i_{1}}\times D_{i_{2}}\times\cdots\times D_{i_{k}},

  • a set of soft constraints 𝒞{CU|UV}\mathcal{C}\subseteq\{C_{U}\ |\ U\subseteq V\} where each soft constraint CU𝒞C_{U}\in\mathcal{C} with U={i1,i2,,ik}U=\{i_{1},i_{2},\dots,i_{k}\} and i1<i2<<iki_{1}<i_{2}<\cdots<i_{k}, is a |U||U|-ary relation CUDi1×Di2××DikC_{U}\subseteq D_{i_{1}}\times D_{i_{2}}\times\cdots\times D_{i_{k}}.

The constraint graph of QQ is defined as G=(V,E)G=(V,E) where E={{u,v}|CU𝒞 s.t. {u,v}U}E=\{\{u,v\}\ |\ \exists C_{U}\in\mathcal{C}\cup\mathcal{H}\textrm{ s.t. }\{u,v\}\subseteq U\}. We say that a CSP instance QQ has bounded treewidth if the constraint graph of QQ has bounded treewidth. In binary CSP, every hard and soft relation is a unary or binary relation, and in boolean CSP, the domain of every variable is {0,1}\{0,1\}. We use DD to denote the maximal size of all domains, that is, D=maxuV|Du|D=\max_{u\in V}|D_{u}|.

For a vector z=(z1,z2,,zn)z=(z_{1},z_{2},\ldots,z_{n}) and U={i1,i2,,ik}VU=\{i_{1},i_{2},\dots,i_{k}\}\subseteq V with i1<i2<<iki_{1}<i_{2}<\cdots<i_{k}, we define the projection of zz on UU as z|U=(zi1,zi2,,zik)z|_{U}=(z_{i_{1}},z_{i_{2}},\ldots,z_{i_{k}}). A vector znz\in\mathbb{R}^{n} satisfies the constraint CU𝒞C_{U}\in\mathcal{C}\cup\mathcal{H} if and only if z|UCUz|_{U}\in C_{U}. We say that a vector z=(z1,,zn)z^{\star}=(z^{\star}_{1},\ldots,z^{\star}_{n}) is a feasible assignment for QQ if zD1×D2××Dnz^{\star}\in D_{1}\times D_{2}\times\ldots\times D_{n} and zz^{\star} satisfies every hard constraint CC\in\mathcal{H}. For a given feasible assignment zz^{\star} we define an extended feasible assignment ex(z)=(z,h)n+|𝒞|(z^{\star})=(z^{\star},h^{\star})\in\mathbb{R}^{n+|\mathcal{C}|} as follows: the coordinates of hh^{\star} are indexed by the soft constraints from 𝒞\mathcal{C} (to be more precise: by the subsets UU of VV used as lower indices of the soft constraints) and for each CU𝒞C_{U}\in\mathcal{C}, we have hU=1h^{\star}_{U}=1 if and only if z|UCUz^{\star}|_{U}\in C_{U}, and hU=0h^{\star}_{U}=0 otherwise. We denote by (Q)\mathcal{F}(Q) the set of all feasible assignments for QQ, by ex(Q)={\mathcal{F}^{ex}(Q)=\{ex(z)|z(Q)}(z^{\star})\ |\ z^{\star}\in\mathcal{F}(Q)\} the set of all extended feasible assignments for QQ. For every instance QQ we define two polytopes: CSP(Q)CSP(Q) is the convex hull of ex(Q)\mathcal{F}^{ex}(Q) and CSP(Q)CSP^{\prime}(Q) the convex hull of (Q)\mathcal{F}(Q). We also define three trivial linear projections:

  • projV(z,h)=z\textrm{proj}_{V}(z,h)=z,       projE(z,h)=h\textrm{proj}_{E}(z,h)=h,       projid(z,h)=(z,h)\textrm{proj}_{id}(z,h)=(z,h)

where znz\in\mathbb{R}^{n} and h|𝒞|h\in\mathbb{R}^{|\mathcal{C}|}, and observe that projV(CSP(Q))=CSP(Q)\textrm{proj}_{V}(CSP(Q))=CSP^{\prime}(Q).

In the decision version of CSP, the set 𝒞\mathcal{C} of soft constraints is empty and the task is to decide whether there exists a feasible assignment. In the maximization (minimization, resp.) version of the problem, the task is to find a feasible assignment that maximizes (minimizes, resp.) the number of satisfied (unsatisfied, resp.) soft constraints. Note that there is no difference between maximization and minimization versions of the problem with respect to optimal solutions but the two versions differ significantly from an approximation perspective.

In the weighted version of CSP we are also given a weight function w:𝒞w:\mathcal{C}\rightarrow\mathbb{R} that specifies for each soft constraint C𝒞C\in\mathcal{C} its weight w(C)w(C). The goal is to find a feasible assignment that maximizes (minimizes, resp.) the total weight of satisfied (unsatisfied, resp.) constraints. The unweighted version of CSP is equivalent to the weighted version with w(C)=1w(C)=1 for all C𝒞C\in\mathcal{C}.

Even more generally, the relations in the soft constraints can be replaced by bounded real valued payoff functions: a soft constraint CU𝒞C_{U}\in\cal C with U={i1,i2,,ik}U=\{i_{1},i_{2},\dots,i_{k}\} is not a |U||U|-ary relation but a function w:Di1×Di2××Dikw:D_{i_{1}}\times D_{i_{2}}\times\ldots\times D_{i_{k}}\rightarrow\mathbb{R} and the payoff of the soft constraint CUC_{U} for a feasible assignment zz^{\star} is w(z|U)w(z^{\star}|_{U}); the objective is to maximize (minimize, resp.) the total payoff. For the sake of simplicity of the presentation we do not consider the problem in this generality although the techniques used in this paper apply in the general setting as well.

For notions related to the treewidth of a graph, we stick to the standard terminology as given in the book by Kloks [10]).

1.2 Related Work

CSP for graphs of bounded treewidth.

As CSP captures many NP-hard problems, it is a natural problem to identify tractable special cases of CSP. Freuder [7] showed that CSP instances with treewidth bounded by τ\tau can be solved in time O(Dτn)O(D^{\tau}n). Later, Grohe et al. [8] proved that, assuming FPTW[1]FPT\not=W[1], this is essentially the only nontrivial class of graphs for which CSP is solvable in polynomial time (cf. Marx [12]).

Describing the polytope of CSP solutions by the means of linear programming, for instances of bounded treewidth, is not a new idea. In 2007, Sellmann et al. published a paper [18] in which they described a linear program that was supposed to define the convex hull of all feasible solutions of a binary CSP when the constraint graph is a tree. They also provided a procedure to convert a given CSP instance with bounded treewidth into one whose constraint graph is a tree, at the cost of blowing up the number of variables and constraints by a function of the treewidth. Unfortunately, there was a substantial bug in their proof and one of the main theorems in the paper does not even hold [17].

The paper [18] also implicitely includes this folklore result: if the constraint graph has treewidth at most τ\tau, then CSP can be solved by τ\tau levels of the Sherali-Adams hierarchy. The resulting formulation is of size 𝒪(nτ)\mathcal{O}(n^{\tau}) while our approach yields size 𝒪(Dτn)\mathcal{O}(D^{\tau}n).

CSP for general graphs.

Chan et al. [4] study the extent to which linear programming relaxation can be used in dealing with approximating CSP. They show that polynomial-sized LPs are exactly as powerful as LPs obtained from a constant number of rounds of the Sherali-Adams hierarchy. They also prove integrality gaps for polynomial-sized LPs for some CSP.

Raghavendra [13] shows that under the Unique Games Conjecture, a certain simple SDP relaxation achieves the best approximation ratio for every CSP. In a follow up paper, Raghavendra and Steurer [14] describe an efficient rounding scheme that achieves the integrality gap of the simple SDP relaxation, and, in another paper [15], they show unconditionally that the integrality gap of this SDP relaxation cannot be reduced by Sherali-Adams hierarchies.

Other related results.

Buchanan and Butenko [3] provide an extended formulation for the independent set problem, a special case of CSP, that has size O(2τn)O(2^{\tau}n) where τ\tau denotes the treewidth of the given graph. Our results can be viewed as a generalization of this result: the size of our formulation, when applied to the independent set problem, is also O(2τn)O(2^{\tau}n).

In a recent work, Bienstock and Munoz [2] define a class of so called general binary optimization problems which are essentially weighted boolean CSP problems, and for instances of treewidth τ\tau provide an LP formulation of size O(2τn)O(2^{\tau}n). Again, this is a special case of our result in this paper. It is worth mentioning at this point that every CSP instance can be transformed into a boolean CSP instance; however, the standard transformation results in a substantial increase (in some cases even Ω(D)\Omega(D)) of the treewidth of the constraint graph.

1.3 New Results

Our main result is summarized as the following theorem.

Theorem 1.1

For every instance Q=(V,𝒟,,𝒞)Q=(V,\mathcal{D},\mathcal{H},\mathcal{C}) of CSP, there exists an extended formulation P(Q)P(Q) of CSP(Q)CSP(Q) and CSP(Q)CSP^{\prime}(Q) of size 𝒪(Dτn)\mathcal{O}(D^{\tau}n) where τ\tau is the treewidth of QQ; moreover, the corresponding LP can be constructed in time 𝒪(Dτn)\mathcal{O}(D^{\tau}n).

As a corollary we obtain upper bounds on the extension complexity for several NP-hard problems on the class of graphs with bounded treewidth; as far as we know, these results have not been known.

2 CSP Polytope

2.1 Integer Linear Programming Formulation

We start by introducing the terms and notation that we use throughout this section. We assume that Q=(V,𝒟,,𝒞)Q=(V,\mathcal{D},\mathcal{H},\mathcal{C}) is a given instance of CSP. For every subset WVW\subseteq V we define the set of all configurations of WW as

𝒦(W)={(α1,,αn)|CU(UWα|UCU) and iUαi=λ}\mathcal{K}(W)=\{(\alpha_{1},\dots,\alpha_{n})\ |\ \forall C_{U}\in\mathcal{H}\ (U\subseteq W\rightarrow\alpha|_{U}\in C_{U})\mbox{ and }\forall i\not\in U\ \alpha_{i}=\lambda\}\

where λ\lambda is a symbol not appearing in any of the domains DuD_{u}, uVu\in V. For a configuration K𝒦(U)K\in\mathcal{K}(U) and vVv\in V, we use the notation K(v)K(v) to refer to the vv-th element of KK. Also, for a configuration K𝒦(U)K\in\mathcal{K}(U), vVUv\in V\setminus U and αDv\alpha\in D_{v}, we use the notation K[vα]K[v\leftarrow\alpha] to denote the configuration K𝒦(U{v})K^{\prime}\in\mathcal{K}(U\cup\{v\}) such that K(v)=αK^{\prime}(v)=\alpha and K(u)=K(u)K^{\prime}(u)=K(u) for every uvu\neq v.

For an nn-dimensional vector K=(α1,,αn)K=(\alpha_{1},\dots,\alpha_{n}) and a subset of variables UVU\subseteq V we denote by KUK\hskip-2.84526pt\upharpoonright_{U} the restriction of KK to UU that is defined as an nn-dimensional vector with KU(i)=K(i)K\hskip-2.84526pt\upharpoonright_{U}(i)=K(i) for iUi\in U and KU(i)=λK\hskip-2.84526pt\upharpoonright_{U}(i)=\lambda for iUi\not\in U (i.e., we set to λ\lambda all coordinates of KK outside of UU). We denote by Λ\Lambda the configuration (λ,,λ)𝒦()(\lambda,\dots,\lambda)\in\mathcal{K}(\emptyset); note that for αDv\alpha\in D_{v}, Λ[vα]\Lambda[v\leftarrow\alpha] is the configuration from 𝒦({v})\mathcal{K}(\{v\}) with exactly one non-λ\lambda element, namely the vv-th element, equaling α\alpha.

In our linear program, for every index vVv\in V and every iDvi\in D_{v}, we introduce a binary variable yviy_{v}^{i}. The task of the variable yviy_{v}^{i} is to encode the value of the CSP-variable zvz_{v}: the variable yviy_{v}^{i} is set to one if and only if zv=iz_{v}=i. Since in every solution each variable assumes a unique value, we enforce the constraint iD(v)yvi=1\sum_{i\in D(v)}y_{v}^{i}=1 for each vVv\in V.

For every configuration KU:CU𝒞𝒦(U)K\in\bigcup_{U:C_{U}\in\mathcal{C}\cup\mathcal{H}}\mathcal{K}(U) we introduce a binary variable g(K)g(K). The intended meaning of the variable g(K)g(K), for K𝒦(U)K\in\mathcal{K}(U) and UVU\subseteq V, is to provide information about the values of the CSP-variables zuz_{u} for uUu\in U in the following way: g(K)=1g(K)=1 if and only if for every uUu\in U, zu=K(u)z_{u}=K(u). To ensure consistency between the yy and gg variables, for every CU𝒞C_{U}\in\mathcal{C}\cup\mathcal{H} and for every vUv\in U, we enforce the constraint K𝒦(U):K(v)=ig(K)=yvi\sum_{K\in\mathcal{K}(U):K(v)=i}g(K)=y_{v}^{i}. Note that for binary CSP, the gg variables capture the values of CSP-variables zz for pairs of elements from VV that correspond to edges of the constraint graph.

Relaxing the integrality constraints we obtain the following initial LP relaxation of the CSP problem Q=(V,𝒟,,𝒞)Q=(V,\mathcal{D},\mathcal{H},\mathcal{C}):

iD(v)yvi\displaystyle\sum_{i\in D(v)}y_{v}^{i} =1\displaystyle=1 vV\displaystyle\forall v\in V (1)
K𝒦(U):K(v)=ig(K)\displaystyle\sum_{K\in\mathcal{K}(U):K(v)=i}g(K) =yvi\displaystyle=y_{v}^{i} CU𝒞vUiD(v)\displaystyle\forall C_{U}\in\mathcal{C}\cup\mathcal{H}\ \forall v\in U\ \forall i\in D(v) (2)
0𝒚,𝒈\displaystyle 0\leq\bm{y},\bm{g} 1\displaystyle\leq 1 (3)

Note that there is a one to one correspondence between the (extended) feasible assignments of QQ and integral solutions of (1) - (3); from now on we denote by proj1\textrm{proj}_{1} the linear projection of the convex hull of integral solutions of (1) - (3) to CSP(Q)CSP(Q). Also observe that the total weight of CSP-constraints satisfied by an integral vector (𝒚,𝒈)(\bm{y},\bm{g}) satisfying (1) - (3) is111In the case of general payoff functions, the total weight is given by CU𝒞K𝒦(U):K|UCUw(K|U)g(K)\sum_{C_{U}\in\mathcal{C}}\sum_{K\in\mathcal{K}(U):K|_{U}\in C_{U}}w(K|_{U})g(K)

CU𝒞w(CU)K𝒦(U):K|UCUg(K).\displaystyle\sum_{C_{U}\in\mathcal{C}}w(C_{U})\sum_{K\in\mathcal{K}(U):K|_{U}\in C_{U}}g(K)\ .

Unfortunately, even for CSP problems whose constraint graph is series-parallel, the polytope given by the LP (1) - (3) is not integral (consider, e.g., the instance of CSP corresponding to the independent set problem on K3K_{3}). The weakness of the formulation is that no global consistency among the 𝒚\bm{y} variables is guaranteed. To strengthen the relaxation, we introduce new variables and constraints derived from a tree decomposition of the constraint graph of QQ.

2.2 Extended Formulation

Here we describe, for every CSP instance Q=(V,𝒟,,𝒞)Q=(V,\mathcal{D},\mathcal{H},\mathcal{C}), a polytope P(Q)P(Q), and in the next subsection we prove that P(Q)P(Q) is an extended formulation of CSP(Q)CSP(Q) and CSP(Q)CSP^{\prime}(Q). The set of variables in the given LP description of P(Q)P(Q) is substantially different from the set of variables used in the LP (1) - (3), and the set of new constraints is completely different from the the set of constraints in the LP (1) - (3). Whereas in the previous subsection, there is (roughly) a variable g(K)g(K) for every feasible assignment of every subset of CSP variables corresponding to a soft or hard constraint, here we have a variable for every feasible assignment of every subset of CSP variables corresponding to a bag in a given tree decomposition of the constraint graph. Nevertheless, as we show after defining P(Q)P(Q), there exists a simple linear projection of P(Q)P(Q) to the convex hull of all integral points in the polytope given by the LP (1) - (3).

Let T=(VT,ET)T=(V_{T},E_{T}) be a fixed nice tree decomposition [10] of the constraint graph of QQ and for every node aVTa\in V_{T}, let B(a)VB(a)\subseteq V denote the corresponding bag. Let ={B(a)|aVT}\mathcal{B}=\{B(a)\ |\ a\in V_{T}\} denote the set of all bags of TT. Let 𝒦=B𝒦(B)\mathcal{K}_{\mathcal{B}}=\bigcup_{B\in\mathcal{B}}\mathcal{K}(B) be the set of all configurations of all bags in TT. We use VIVTV_{I}\subseteq V_{T} to denote the subset of all introduce nodes in TT and VFVTV_{F}\subseteq V_{T} to denote the subset of all forget nodes in TT.

For every configuration K𝒦K\in\mathcal{K}_{\mathcal{B}} we introduce a binary variable f(K)f(K). As in the previous subsection, the intended meaning of the variable K𝒦(B)K\in\mathcal{K}(B), for BB\in\mathcal{B}, is to provide information about the values of the CSP-variables zuz_{u} for uBu\in B in the following way: f(K)=1f(K)=1 if and only if for every uBu\in B, zu=K(u)z_{u}=K(u). To ensure consistency among variables indexed by the configurations of the same bag, namely to ensure that for every BB\in\mathcal{B} there exists exactly one configuration K𝒦(B)K\in\mathcal{K}(B) with f(K)=1f(K)=1, we introduce for every BB\in\mathcal{B} the LP constraint K𝒦(B)f(K)=1\sum_{K\in\mathcal{K}(B)}f(K)=1.

For every introduce node cVTc\in V_{T} with a child bVTb\in V_{T} and for every configuration K𝒦(B(b))K\in\mathcal{K}(B(b)) we have the constraint K𝒦(B(c)):KB(b)=Kf(K)=f(K)\sum_{K^{\prime}\in\mathcal{K}(B(c)):K^{\prime}\ \hskip-2.84526pt\upharpoonright_{B(b)}=K}f(K^{\prime})=f(K), and symmetrically, for every forget node cVTc\in V_{T} with a child bVTb\in V_{T} and for every configuration K𝒦(B(c))K\in\mathcal{K}(B(c)) we have the constraint K𝒦(B(b)):KB(c)=Kf(K)=f(K)\sum_{K^{\prime}\in\mathcal{K}(B(b)):K^{\prime}\ \hskip-2.84526pt\upharpoonright_{B(c)}=K}f(K^{\prime})=f(K).

Relaxing the integrality constraints and putting all these additional constraints together, we obtain:

K𝒦(B)f(K)\displaystyle\sum_{K\in\mathcal{K}(B)}f(K) =1\displaystyle=1 B\displaystyle\forall B\in\mathcal{B} (4)
K𝒦(B(c)):KB(b)=Kf(K)\displaystyle\sum_{K^{\prime}\in\mathcal{K}(B(c)):K^{\prime}\ \hskip-2.84526pt\upharpoonright_{B(b)}=K}f(K^{\prime}) =f(K)\displaystyle=f(K) cVI,K𝒦(B(b)) where b is\displaystyle\forall c\in V_{I},\forall K\in\mathcal{K}(B(b))\mbox{ where $b$ is } (5)
the only child of cc
K𝒦(B(b)):KB(c)=Kf(K)\displaystyle\sum_{K^{\prime}\in\mathcal{K}(B(b)):K^{\prime}\ \hskip-2.84526pt\upharpoonright_{B(c)}=K}f(K^{\prime}) =f(K)\displaystyle=f(K) cVF,K𝒦(B(c)) where b is\displaystyle\forall c\in V_{F},\forall K\in\mathcal{K}(B(c))\mbox{ where $b$ is } (6)
the only child of cc
0𝒇\displaystyle 0\leq\bm{f} 1\displaystyle\leq 1 (7)

For the given binary CSP instance QQ, we denote the polytope associated with the LP (4) - (7), as P(Q)P(Q).

Consider now a vector 𝒇P(Q)\bm{f}\in P(Q) and the following set of linear equations:

yvi\displaystyle y_{v}^{i} =K𝒦(B):K(v)=if(K)\displaystyle=\sum_{{K\in\mathcal{K}(B):K(v)=i}}f(K) B,vB,iDv\displaystyle\forall B\in\mathcal{B},\forall v\in B,\forall i\in D_{v} (8)
g(K)\displaystyle g(K) =K𝒦(B):KU=Kf(K)\displaystyle=\sum_{K^{\prime}\in\mathcal{K}(B):K^{\prime}\ \hskip-2.84526pt\upharpoonright_{U}=K}f(K^{\prime}) B,CU𝒞 s.t. UB,K(U)\displaystyle\forall B\in\mathcal{B},\forall C_{U}\in\mathcal{C}\cup\mathcal{H}\mbox{ s.t. }U\subseteq B,\forall K\in\mathcal{B}(U) (9)

It is just a technical exercise to check that for a given 𝒇P(Q)\bm{f}\in P(Q), there always exists a unique solution (𝒚,𝒈)(\bm{y},\bm{g}) of this LP and that the unique (𝒚,𝒈)(\bm{y},\bm{g}) is a linear projection of 𝒇\bm{f}. Moreover, such a vector (𝒚,𝒈)(\bm{y},\bm{g}) also satisfies the LP constraints (1) - (3). The point is that there exists a linear projection of P(Q)P(Q) into the polytope defined by the LP (1) - (3); moreover, an integral point from P(Q)P(Q) is mapped on an integral point. From now on we denote this projection proj2\textrm{proj}_{2}.

2.3 Proof of Theorem 1.1

As in the previous subsections, we assume that Q=(V,𝒟,,𝒞)Q=(V,\mathcal{D},\mathcal{H},\mathcal{C}) is a given instance of CSP, G=(V,E)G=(V,E) is the constraint graph of QQ and T=(VT,ET)T=(V_{T},E_{T}) a fixed nice tree decomposition of GG. We start by introducing several notions that will help us dealing with tree decompositions and our linear program.

For a node aVTa\in V_{T}, let T(a)=(Va,Ea)T(a)=(V_{a},E_{a}) be the subtree of TT rooted in aa; the configurations relevant to T(a)T(a) are those in the set (a)=bVa𝒦(B(b))\mathcal{R}(a)=\bigcup_{b\in V_{a}}\mathcal{K}(B(b)), and the variables relevant to T(a)T(a) are those f(K)f(K) for which K(a)K\in\mathcal{R}(a). For succinctness of notation, we denote the projection 𝒇|(a)\bm{f}|_{\mathcal{R}(a)} of the vector 𝒇\bm{f} on the set of variables relevant to T(a)T(a) also by 𝒇|a\bm{f}|_{a}. The constraints relevant to T(a)T(a) are those containing only the variables relevant to T(a)T(a). We say that a vector I{0,1}(a)I\in\{0,1\}^{\mathcal{R}(a)} agrees with the configuration K(a)K\in\mathcal{R}(a) if I(K)=1I(K)=1.

Let 𝒇\bm{f} be a fixed solution of the LP (4) - (7) that corresponds to a vertex of the polytope P(Q)P(Q). Our main tool is the following lemma.

Lemma 1

For every node bVTb\in V_{T}, there exist a positive integer MM and binary vectors I1,I2,,IM{0,1}(b)I_{1},I_{2},\dots,I_{M}\in\{0,1\}^{\mathcal{R}(b)}, some possibly identical, such that

  • \spadesuit

    every IiI_{i} satisfies the constraints relevant to T(b)T(b),

  • \clubsuit

    𝒇|b=1Mi=1MIi\bm{f}|_{b}=\frac{1}{M}\sum_{i=1}^{M}I_{i}.

Proof

By induction. We start in the leaves of TT and proceed in a bottom-up fashion.

Base case.

Assume that bVTb\in V_{T} is a leaf of the nice decomposition tree TT. By definition of a nice tree decomposition, the bag B(b)B(b) consists of a single vertex, say a vertex vVv\in V. The only variables relevant to T(b)T(b) are f(K)f(K) for all K𝒦(B(b))=jD(v)Λ[vj]K\in\mathcal{K}(B(b))=\bigcup_{j\in D(v)}\Lambda[v\leftarrow j], and the only relevant constraints are those of the type (4) and (7).

Let MM^{\prime}\in\mathbb{N} be such that an MM^{\prime}-multiple of every relevant variable is integral; as 𝒇\bm{f} is a solution corresponding to a vertex of the polytope P(Q)P(Q), all the variables are rational which guarantees that such an MM^{\prime} exists. For every jDvj\in D_{v} we define an integral vector IjI_{j} such that Ij(Λ[vj])=1I_{j}(\Lambda[v\leftarrow j])=1 and Ij(Λ[vi])=0I_{j}(\Lambda[v\leftarrow i])=0 for every iji\neq j.

The vector IjI_{j} will appear with multiplicity MyvjM^{\prime}\cdot y_{v}^{j} among the integral solutions I1,,IMI_{1},\ldots,I_{M^{\prime}} for GG^{\prime}. Then, obviously, both properties \spadesuit and \clubsuit are satisfied.

Inductive step.

Consider an internal node cVTc\in V_{T} of the nice decomposition tree TT. We distinguish three cases: cc is a join node, cc is an introduce node and cc is a forget node.

Join node. Assume that the two children of the join node cc are aa and bb. Recall that B(a)=B(b)=B(c)B(a)=B(b)=B(c). By the inductive assumption, there exist integers MM and MM^{\prime} and integral vectors I1,,IM{0,1}(a)I_{1},\ldots,I_{M}\in\{0,1\}^{\mathcal{R}(a)}, each of them satisfying the relevant constraints for T(a)T(a) and such that 𝒇|a=1Mi=1MIi\bm{f}|_{a}=\frac{1}{M}\sum_{i=1}^{M}I_{i}, and integral vectors J1,,JM{0,1}(b)J_{1},\ldots,J_{M^{\prime}}\in\{0,1\}^{\mathcal{R}(b)}, each of them satisfying the relevant constraints for T(b)T(b) and such that 𝒇|b=1Mi=1MJi\bm{f}|_{b}=\frac{1}{M^{\prime}}\sum_{i=1}^{M^{\prime}}J_{i}.

Two vectors IiI_{i} and JjJ_{j} that agree with a given configuration K𝒦(B(c))K\in\mathcal{K}(B(c)) can be easily merged into an integral vector L{0,1}(c)L\in\{0,1\}^{\mathcal{R}(c)} that satisfies L|a=IiL|_{a}=I_{i} and L|b=JjL|_{b}=J_{j}; as the set of all constraints relevant to T(c)T(c) is the union of the constraints relevant to T(a)T(a) and the constraints relevant to T(b)T(b), the vector LL satisfies also all the constraints relevant to T(c)T(c).

For simplicity we assume, without loss of generality, that M=MM=M^{\prime}. Then, by the property \clubsuit and since B(a)=B(b)=B(c)B(a)=B(b)=B(c), for every configuration K𝒦(B(c))K\in\mathcal{K}(B(c)), the number of vectors IiI_{i} that agree with KK is equal to the number of vectors JjJ_{j} that agree with KK, namely Mf(K)M\cdot f(K). Thus, it is possible to match the vectors IiI_{i} and JjJ_{j} one to one in such a way that both vectors in each pair agree with the same configuration; let L1,L2,,LML_{1},L_{2},\ldots,L_{M} denote the result of their merging as described above. Then the vectors LiL_{i} satisfy the property \spadesuit as explained in the previous paragraph, and by construction they also satisfy the property \clubsuit.

Introduce node. Assume that the only child of the introduce node cc is a node bb and B(c)=B(b){v}B(c)=B(b)\cup\{v\}. By the inductive assumption, there exists integer MM and integral vectors I1,,IM{0,1}(b)I_{1},\ldots,I_{M}\in\{0,1\}^{\mathcal{R}(b)}, each of them satisfying the relevant constraints for T(b)T(b) and such that 𝒇|b=1Mi=1MIi\bm{f}|_{b}=\frac{1}{M}\sum_{i=1}^{M}I_{i}. Without loss of generality we assume that for every variable relevant to T(c)T(c), its MM-multiple is integral. We partition the vectors I1,,IMI_{1},\ldots,I_{M} into several groups indexed by the configurations from 𝒦(B(b))\mathcal{K}(B(b)): the group ZKZ_{K}, for K𝒦(B(b))K\in\mathcal{K}(B(b)), consists exactly of those vectors IiI_{i} that agree with KK.

Consider a fixed configuration K𝒦(B(b))K\in\mathcal{K}(B(b)) and the corresponding group ZKZ_{K}. Note that the size of this group is Mf(K)M\cdot f(K). We further partition the group ZKZ_{K} into at most |Dv||D_{v}| subgroups ZKZ_{K^{\prime}}, where K=K[vj]K^{\prime}=K[v\leftarrow j], for every jDvj\in D_{v} satisfying K[vj]𝒦(B(c))K[v\leftarrow j]\in\mathcal{K}(B(c)), in such a way that ZKZ_{K^{\prime}} contains exactly Mf(K)M\cdot f(K^{\prime}) vectors (it does not matter which ones); the LP constraint (5) makes this possible. Then, for every jDvj\in D_{v}, we create from every vector IZK[vj]I\in Z_{K[v\leftarrow j]} a new integral vector JIJ_{I} in the following way:

  • for every K¯(b)\bar{K}\in\mathcal{R}(b), JI(K¯)=I(K¯)J_{I}(\bar{K})=I(\bar{K}); this guarantees JI|b=IJ_{I}|_{b}=I,

  • JI(K[vj])=1J_{I}(K[v\leftarrow j])=1,

  • for every iDvi\in D_{v}, iji\not=j, JI(K[vi])=0J_{I}(K[v\leftarrow i])=0.

Obviously, the new vectors JIJ_{I} satisfy all constraints relevant to T(b)T(b), and it is easy to check that they satisfy all constraints relevant to T(c)T(c) as well, given the definitions above. Moreover, the definitions above imply that the vectors JIJ_{I} satisfy the property \clubsuit.

Forget node. Assume that the only child of the forget node cc is a node bb, B(c)=B(b){v}B(c)=B(b)\setminus\{v\}. This case is symmetric to the previous one in that instead of splitting the groups ZKZ_{K} into smaller groups ZKZ_{K^{\prime}}, we merge them into bigger ZKZ_{K^{\prime}}.

By the inductive assumption, there exists an integer MM and integral vectors I1,,IM{0,1}(b)I_{1},\ldots,I_{M}\in\{0,1\}^{\mathcal{R}(b)}, each of them satisfying the relevant constraints for T(b)T(b) and such that 𝒇|b=1Mi=1MIi\bm{f}|_{b}=\frac{1}{M}\sum_{i=1}^{M}I_{i}. Without loss of generality we assume that for every variable relevant to T(c)T(c), its MM-multiple is integral. We partition the vectors I1,,IMI_{1},\ldots,I_{M} into several groups indexed by the configurations from 𝒦(B(b))\mathcal{K}(B(b)): the group ZKZ_{K}, for K𝒦(B(b))K\in\mathcal{K}(B(b)), consists exactly of those vectors IiI_{i} that agree with KK. Note that the size of ZKZ_{K} is Mf(K)M\cdot f(K).

For every K𝒦(B(c))K^{\prime}\in\mathcal{K}(B(c)) we create a bigger group group ZKZ_{K^{\prime}} by merging |Dv||D_{v}| of the groups ZKZ_{K}, namely those satisfying K|B(c)=KK|_{B(c)}=K^{\prime}. By the LP constraint (6), the new group ZKZ_{K^{\prime}} contains exactly Mf(K)M\cdot f(K^{\prime}) vectors. For every K𝒦(B(c))K^{\prime}\in\mathcal{K}(B(c)), we create from every vector IZKI\in Z_{K^{\prime}} a new integral vector JIJ_{I} in the following way:

  • for every K¯(b)\bar{K}\in\mathcal{R}(b), JI(K¯)=I(K¯)J_{I}(\bar{K})=I(\bar{K}).

If 𝒦(B(c))(b)\mathcal{K}(B(c))\subseteq\mathcal{R}(b), there is nothing more to do. Otherwise we further define

  • JI(K)=1J_{I}(K^{\prime})=1, and for every K^𝒦(B(c))\hat{K}\in\mathcal{K}(B(c)), K^K\hat{K}\not=K^{\prime}, JI(K^)=0J_{I}(\hat{K})=0.

We have to check that the vectors JIJ_{I} satisfy all constraints relevant to T(c)T(c). The only possibly new constraints are those using variables f(K)f(K^{\prime}) for K𝒦(B(c))K^{\prime}\in\mathcal{K}(B(c)) and it is easily seen that they are satisfied, given the definitions above. Also, the definitions above imply that the vectors JKJ_{K^{\prime}} satisfy the property \clubsuit. ∎

By applying Lemma 1 to the whole tree TT, that is, to the subtree rooted in the root of TT, we immediately obtain that 𝒇\bm{f} is an integral vector, and, thus, also the corresponding vertex of P(Q)P(Q) is integral. As this holds for every vertex of P(Q)P(Q), we conclude that P(Q)P(Q) is an integral polytope.

Considering the notes at the ends of the previous two subsections, we also conclude that CSP(Q)=proj1(proj2(P(Q))CSP(Q)=\textrm{proj}_{1}(\textrm{proj}_{2}(P(Q)) and CSP(Q)=projV(CSP(Q))CSP^{\prime}(Q)=\textrm{proj}_{V}(CSP(Q)).

To complete the proof of Theorem 1.1, we observe that the number of variables and constraints in the LP (4) - (7) is 𝒪(Dτn)\mathcal{O}(D^{\tau}n). ∎

3 Applications

The purpose of this section is to make explicit the extension complexity upper bounds given in Theorem 1.1 for several well known graph problems. We find it interesting that the attained extension complexity upper bounds meet the best possible (assuming Strong ETH) time complexity lower bounds, given by Lokshtanov et al. [11]; the only exception is the Multiway Cut problem. To state our results, we use for each problem the following template:

 

Problem name

Projection

Extension complexity

Time complexity

Instance:   …

Solution:   …

CSP formulation: VV, 𝒟\mathcal{D}, \mathcal{H}, 𝒞\mathcal{C}. CSP version: Decision / Max / Min

where Projection is the name of the linear projection that yields the natural polytope of the problem QQ from the CSP(Q)CSP(Q) polytope (or from the P(Q)P(Q) polytope, in case of the OCT problem). We use the notation [n]={1,,n}[n]=\{1,\dots,n\}.

 

Coloring / Chromatic Number [1]

projV\textrm{proj}_{V}

𝒪(\mathcal{O}(qτnq^{\tau}n))

Θ(\Theta(qτnq^{\tau}n))

Instance: Graph G=(V,E)G=(V,E), set of colors [q][q]

Solution: A coloring of GG with qq colors with no monochromatic edges.

CSP formulation: V=[n]V=[n], Dv=[q]D_{v}=[q] for every vVv\in V, Huv={(i,j)|iDu,jDv,ij}H_{uv}=\{(i,j)\ |\ i\in D_{u},j\in D_{v},i\neq j\} for every uvEuv\in E, 𝒞=\mathcal{C}=\emptyset. Decision

Comment: Note that Chromatic Number χ(G)\chi(G) of GG is always upper bounded by τ+1\tau+1 since graphs of bounded treewidth are τ\tau-degenerate and thus (τ+1)(\tau+1)-colorable. Thus, if the goal is to determine χ(G)\chi(G), it suffice to find the smallest qq such that CSP(Q)CSP(Q) is non-empty.

 

List-HH-Coloring / List Homomorphism [6]

projV\textrm{proj}_{V}

𝒪(\mathcal{O}(LτnL^{\tau}n))

Θ(\Theta(LτnL^{\tau}n))

Instance: Graph G=(V,E)G=(V,E), graph H=(VH,EH)H=(V_{H},E_{H}) possibly containing loops, and for every vertex vVv\in V a set L(v)VHL(v)\subseteq V_{H}. (We denote L=maxvV|L(v)|L=\max_{v\in V}|L(v)|)

Solution: A mapping f:VVHf:V\rightarrow V_{H} such that uvE\forall uv\in E it holds that f(u)f(v)EHf(u)f(v)\in E_{H} and f(v)L(v)f(v)\in L(v) for every vVv\in V.

CSP formulation: V=[n]V=[n], Dv=L(v)D_{v}=L(v) for every vVv\in V, Huv={(i,j)|iDu,jDv,ijEH}H_{uv}=\{(i,j)\ |\ i\in D_{u},j\in D_{v},ij\in E_{H}\} for every uvEuv\in E, 𝒞=\mathcal{C}=\emptyset. Decision

Comment: Note that the problems List Coloring, Precoloring Extension and HH-Coloring (or Graph Homomorphism) are all special cases of this problem. The lower bound given by Lokshtanov et al. [11] applies to all of them since Coloring is a special case of each of them.

 

Unique Games [9]

projid\textrm{proj}_{id}

𝒪(tτn)\mathcal{O}(t^{\tau}n)

Instance: Graph G=(V,E)G=(V,E), an integer tt\in\mathbb{N}, a permutation πe\pi_{e} of order tt for every edge eEe\in E.

Solution: A mapping :V[t]\ell:V\rightarrow[t] such that the number of edges uvEuv\in E with πuv((u))=(v)\pi_{uv}(\ell(u))=\ell(v) is maximized.

CSP formulation: V=[n]V=[n], Dv=[t]D_{v}=[t] for every vVv\in V, =\mathcal{H}=\emptyset, Cuv={(i,πuv(i))|iDu}C_{uv}=\{(i,\pi_{uv}(i))\ |\ i\in D_{u}\} for every edge uvEuv\in E. Max

Comment: The decision variant of this problem is not interesting as it is trivially solvable in polynomial time.

 

Multiway Cut [1]

projE\textrm{proj}_{E}

𝒪(tτn)\mathcal{O}(t^{\tau}n)

𝒪(tτn)\mathcal{O}(t^{\tau}n)

Instance: Graph G=(V,E)G=(V,E), an integer tt\in\mathbb{N} and tt vertices s1,,stVs_{1},\dots,s_{t}\in V

Solution: A partition of VV into sets V1,,VtV_{1},\dots,V_{t} such that for every ii we have siVis_{i}\in V_{i} and the total number of edges between ViV_{i} and VjV_{j} for iji\neq j is minimized.

CSP formulation: V=[n]V=[n], Dv=[t]D_{v}=[t] for every vVv\in V, =\mathcal{H}=\emptyset, Cuv={(i,i)|i[n]}C_{uv}=\{(i,i)\ |\ i\in[n]\} for every edge uvEuv\in E. Min

Comment: Setting zv=iz_{v}=i models vertex vv belonging to the set ViV_{i}. Not satisfying the constraint CuvC_{uv} means that the edge uvuv belongs to the multiway cut.

 

Max Cut [1]

projE\textrm{proj}_{E}

𝒪(\mathcal{O}(2τn2^{\tau}n))

Θ(\Theta(2τn2^{\tau}n))

Instance: Graph G=(V,E)G=(V,E)

Solution: A partition of vertices into two sets V1,V2V_{1},V_{2} such that the number of edges between V1V_{1} and V2V_{2} is maximized.

CSP formulation: V=[n]V=[n], Dv={0,1}D_{v}=\{0,1\} for every vVv\in V, =\mathcal{H}=\emptyset, Cuv={(1,0),(0,1)}C_{uv}=\{(1,0),(0,1)\} for every edge uvEuv\in E. Max

Comment: The values 0,10,1 model the vertex belonging to the set V1V_{1} or V2V_{2}. If we replace maximization by minimization, the problem becomes Edge Bipartization (aka Edge OCT) problem which is a parametric dual of Max Cut.

 

Vertex Cover [1]

projV\textrm{proj}_{V}

𝒪(\mathcal{O}(2τn2^{\tau}n))

Θ(\Theta(2τn2^{\tau}n))

Instance: Graph G=(V,E)G=(V,E)

Solution: A set of vertices CVC\subseteq V of minimal size such that every edge contains a vertex vCv\in C as at least one of its endpoints.

CSP formulation: V=[n]V=[n], Dv={0,1}D_{v}=\{0,1\} for every vVv\in V, Huv={(0,0),(0,1),(1,0)}H_{uv}=\{(0,0),(0,1),(1,0)\} for every edge uvEuv\in E, Cv={1}C_{v}=\{1\}. Min

Comment: The values 0,10,1 model the vertex belonging to CC or VCV\setminus C. If we replace maximization by minimization, the problem becomes Independent Set problem which is a parametric dual of Vertex Cover.

 

Odd Cycle Transversal [11]

projOCT\textrm{proj}_{OCT} \circ proj2\textrm{proj}_{2}

𝒪(3τn)\mathcal{O}(3^{\tau}n)

Θ(3τn)\Theta(3^{\tau}n)

Instance: Graph G=(V,E)G=(V,E)

Solution: A subset of vertices WVW\subseteq V of minimal size such that G[VW]G[V\setminus W] is a bipartite graph.

CSP formulation: V=[n]V=[n], Dv={0,1,2}D_{v}=\{0,1,2\} for every vVv\in V, Huv={0,1,2}2{(0,0),(1,1)}H_{uv}=\{0,1,2\}^{2}\setminus\{(0,0),(1,1)\} for every edge uvEuv\in E, Cv={0,1}C_{v}=\{0,1\} for every vVv\in V. Min

Comment: The values 0,1,20,1,2 model the vertex belonging to either the first or the second partite of a bipartite graph, or the deletion set WW. Satisfying the constraint CvC_{v} corresponds to not putting vv in the deletion set WW. Also known as Vertex Bipartization. The projection projOCT:P(Q){0,1}V\textrm{proj}_{OCT}:P(Q)\rightarrow\{0,1\}^{V} is defined as follows: projOCT(y10,y11,y12,y20,y21,y22,,yn0,yn1,yn2,𝒈)=(y12,y22,,yn2)\textrm{proj}_{OCT}(y_{1}^{0},y_{1}^{1},y_{1}^{2},y_{2}^{0},y_{2}^{1},y_{2}^{2},\ldots,y_{n}^{0},y_{n}^{1},y_{n}^{2},\bm{g})=(y_{1}^{2},y_{2}^{2},\ldots,y_{n}^{2}).

4 Open problems

A natural research direction is to examine more closely the extension complexity for CSP and the specific graph problems on graphs with bounded treewidth, in particular, what are the best possible upper bounds?

Acknowledgments.

The authors thank Hans Raj Tiwary and Jiří Sgall for stimulating discussions.

References

  • [1] G. Ausiello, P. Creczenzi, G. Gambosi, V. Kann, A. Marchetti-Spaccamela, and M. Protasi. Complexity and Approximation; Combinatorial Optimization Problems and Their Approximability Properties. Springer, 1999.
  • [2] D. Bienstock and G. Munoz. LP approximations to mixed-integer polynomial optimization problems. ArXiv e-prints, Jan. 2015.
  • [3] A. Buchanan and S. Butenko. Tight extended formulations for independent set, 2014. Available on Optimization Online.
  • [4] S. O. Chan, J. R. Lee, P. Raghavendra, and D. Steurer. Approximate constraint satisfaction requires large LP relaxations. In Proc. of the 54th Annual IEEE Symposium on Foundations of Computer Science, (FOCS), pages 350–359, 2013.
  • [5] M. Conforti, G. Cornuéjols, and G. Zambelli. Extended formulations in combinatorial optimization. Annals OR, 204(1):97–143, 2013.
  • [6] T. Feder and P. Hell. List homomorphisms to reflexive graphs. J. Comb. Theory, Ser. B, 72(2):236–250, 1998.
  • [7] E. C. Freuder. Complexity of KK-tree structured constraint satisfaction problems. In Proc. of the 8th National Conference on Artificial Intelligence, pages 4–9, 1990.
  • [8] M. Grohe, T. Schwentick, and L. Segoufin. When is the evaluation of conjunctive queries tractable? In Proc. of the 33rd Annual ACM Symposium on Theory of Computing (STOC), pages 657–666, 2001.
  • [9] S. Khot. On the power of unique 2-Prover 1-Round games. In Proc. of the 34th Annual ACM Symposium on Theory of Computing (STOC), pages 767–775, 2002.
  • [10] T. Kloks. Treewidth: Computations and Approximations, volume 842 of Lecture Notes in Computer Science. Springer, 1994.
  • [11] D. Lokshtanov, D. Marx, and S. Saurabh. Known algorithms on graphs on bounded treewidth are probably optimal. In Proc. of the 22nd Annual ACM-SIAM Symposium on Discrete Algorithms, (SODA), pages 777–789, 2011.
  • [12] D. Marx. Can you beat treewidth? Theory of Computing, 6(1):85–112, 2010.
  • [13] P. Raghavendra. Optimal algorithms and inapproximability results for every CSP? In Proc. of the 40th Annual ACM Symposium on Theory of Computing (STOC), pages 245–254, 2008.
  • [14] P. Raghavendra and D. Steurer. How to round any CSP. In Proc. of the 50th Annual IEEE Symposium on Foundations of Computer Science, (FOCS), pages 586–594, 2009.
  • [15] P. Raghavendra and D. Steurer. Integrality gaps for strong SDP relaxations of unique games. In Proc. of the 50th Annual IEEE Symposium on Foundations of Computer Science, (FOCS), pages 575–585, 2009.
  • [16] T. Rothvoß. The matching polytope has exponential extension complexity. In Proc. of the 46th ACM Symposium on Theory of Computing, (STOC), pages 263–272, 2014.
  • [17] M. Sellmann. The polytope of tree-structured binary constraint satisfaction problems. In Proc. of Integration of AI and OR Techniques in Constraint Programming for Combinatorial Optimization Problems (CPAIOR), volume 5015 of Lecture Notes in Computer Science, pages 367–371. Springer, 2008.
  • [18] M. Sellmann, L. Mercier, and D. H. Leventhal. The linear programming polytope of binary constraint problems with bounded tree-width. In Proc. of Integration of AI and OR Techniques in Constraint Programming for Combinatorial Optimization Problems (CPAIOR), volume 4510 of Lecture Notes in Computer Science, pages 275–287. Springer, 2007.