This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deterministic Mincut in Almost-Linear Time

Jason Li
Carnegie Mellon University
Most of this work was done while the author was an intern at Microsoft Research, Redmond.
Abstract

We present a deterministic (global) mincut algorithm for weighted, undirected graphs that runs in m1+o(1)m^{1+o(1)} time, answering an open question of Karger from the 1990s. To obtain our result, we de-randomize the construction of the skeleton graph in Karger’s near-linear time mincut algorithm, which is its only randomized component. In particular, we partially de-randomize the well-known Benczur-Karger graph sparsification technique by random sampling, which we accomplish by the method of pessimistic estimators. Our main technical component is designing an efficient pessimistic estimator to capture the cuts of a graph, which involves harnessing the expander decomposition framework introduced in recent work by Goranci et al. (SODA 2021). As a side-effect, we obtain a structural representation of all approximate mincuts in a graph, which may have future applications.

1 Introduction

The minimum cut of an undirected, weighted graph G=(V,E,w)G=(V,E,w) is a minimum weight subset of edges whose removal disconnects the graph. Finding the mincut of a graph is one of the central problems in combinatorial optimization, dating back to the work of Gomory and Hu [GH61] in 1961 who gave an algorithm to compute the mincut of an nn-vertex graph using n1n-1 max-flow computations. Since then, a large body of research has been devoted to obtaining faster algorithms for this problem. In 1992, Hao and Orlin [HO92] gave a clever amortization of the n1n-1 max-flow computations to match the running time of a single max-flow computation. Using the “push-relabel” max-flow algorithm of Goldberg and Tarjan [GT88], they obtained an overall running time of O(mnlog(n2/m))O(mn\log(n^{2}/m)) on an nn-vertex, mm-edge graph. Around the same time, Nagamochi and Ibaraki [NI92a] (see also [NI92b]) designed an algorithm that bypasses max-flow computations altogether, a technique that was further refined by Stoer and Wagner [SW97] (and independently by Frank in unpublished work). This alternative method yields a running time of O(mn+n2logn)O(mn+n^{2}\log n). Before 2020, these works yielding a running time bound of O~(mn)\widetilde{O}(mn) were the fastest deterministic mincut algorithms for weighted graphs.

Starting with Karger’s contraction algorithm in 1993 [Kar93], a parallel body of work started to emerge in randomized algorithms for the mincut problem. This line of work (see also Karger and Stein [KS96]) eventually culminated in a breakthrough paper by Karger [Kar00] in 1996 that gave an O(mlog3n)O(m\log^{3}n) time Monte Carlo algorithm for the mincut problem. Note that this algorithm comes to within poly-logarithmic factors of the optimal O(m)O(m) running time for this problem. In that paper, Karger asks whether we can also achieve near-linear running time using a deterministic algorithm. Even before Karger’s work, Gabow [Gab95] showed that the mincut can be computed in O(m+λ2nlog(n2/m))O(m+\lambda^{2}n\log(n^{2}/m)) (deterministic) time, where λ\lambda is the value of the mincut (assuming integer weights). Note that this result obtains a near-linear running time if λ\lambda is a constant, but in general, the running time can be exponential. Indeed, for general graphs, Karger’s question remains open after more than 20 years. However, some exciting progress has been reported in recent years for special cases of this problem. In a recent breakthrough, Kawarabayashi and Thorup [KT18] gave the first near-linear time deterministic algorithm for this problem for simple graphs. They obtained a running time of O(mlog12n)O(m\log^{12}n), which was later improved by Henzinger, Rao, and Wang [HRW17] to O(mlog2nloglog2n)O(m\log^{2}n\log\log^{2}n), and then simplified by Saranurak [Sar21] at the cost of m1+o(1)m^{1+o(1)} running time. From a technical perspective, Kawarabayashi and Thorup’s work introduced the idea of using low conductance cuts to find the mincut of the graph, a very powerful idea that we also exploit in this paper.

In 2020, the author, together with Debmalya Panigrahi [LP20], obtained the first improvement to deterministic mincut for weighted graphs since the 1990s, obtaining a running time of O(m1+ϵ)O(m^{1+\epsilon}) plus polylogarithmic calls to a deterministic exact sstt max-flow algorithm. Using the fastest deterministic algorithm for weighted graphs of Goldberg and Rao [GR98], their running time becomes O~(m1.5)\widetilde{O}(m^{1.5}).111In this paper, O~()\widetilde{O}(\cdot) notation hides polylogarithmic factors in nn, the number of vertices of the graph. Their algorithm was inspired by the conductance-based ideas of Kawarabayashi and Thorup and introduced expander decompositions into the scene. While it is believed that a near-linear time algorithm exists for sstt max-flow—which, if deterministic, would imply a near-linear time algorithm for deterministic mincut—the best max-flow algorithms, even for unweighted graphs, is still m4/3+o(1)m^{4/3+o(1)} [LS20]. For the deterministic, weighted case, no improvement since Goldberg-Rao [GR98] is known.

The main result of this paper is a new deterministic algorithm for mincut that does not rely on sstt max-flow computations and achieves a running time of m1+o(1)m^{1+o(1)}, answering Karger’s open question.

Theorem 1.1.

There is a deterministic mincut algorithm for weighted, undirected graphs that runs in m1+o(1)m^{1+o(1)} time.

1.1 Our Techniques

Our approach differs fundamentally from the one in [LP20] that relies on sstt max-flow computations. At a high level, we follow Karger’s approach and essentially de-randomize the single randomized procedure in Karger’s near-linear time mincut algorithm [Kar00], namely the construction of the skeleton graph, which Karger accomplishes through the Benczur-Karger graph sparsification technique by random sampling. We remark that our de-randomization does not recover a full (1+ϵ)(1+\epsilon)-approximate graph sparsifier, but the skeleton graph that we obtain is sufficient to solve the mincut problem.

Let us first briefly review the Benczur-Karger graph sparsification technique, and discuss the difficulties one encounters when trying to de-randomize it. Given a weighted, undirected graph, the sparsification algorithm samples each edge independently with a probability depending on the weight of the edge and the global mincut of the graph, and then re-weights the sampled edge accordingly. In traditional graph sparsification, we require that every cut in the graph has its weight preserved up to a (1+ϵ)(1+\epsilon) factor. There are exponentially many cuts in a graph, so a naive union bound over all cuts does not work. Benczur and Karger’s main insight is to set up a more refined union bound, layering the (exponentially many) cuts in a graph by their weight. They show that for all α1\alpha\geq 1, there are only ncαn^{c\alpha} many cuts in a graph whose weight is roughly α\alpha times the mincut, and each one is preserved up to a (1+ϵ)(1+\epsilon) factor with probability 1ncα1-n^{-c^{\prime}\alpha}, for some constants ccc^{\prime}\gg c. In other words, they establish a union bound layered by the α\alpha-approximate mincuts of a graph, for each α1\alpha\geq 1.

One popular method to de-randomize random sampling algorithms is through pessimistic estimators, which is a generalization of the well-known method of conditional probabilities. For the graph sparsification problem, the method of pessimistic estimators can be implemented as follows. The algorithm considers each edge one by one in some arbitrary order, and decides on the spot whether to keep or discard each edge for the sparsifier. To make this decision, the algorithm maintains a pessimistic estimator, which is a real number in the range [0,1)[0,1) that represents an upper bound on the probability of failure should the remaining undecided edges each be sampled independently at random. In many cases, the pessimistic estimator is exactly the probability upper bound that one derives from analyzing the random sampling algorithm, except conditioned on the edges kept and discarded so far. The algorithm makes the choice—whether to keep or discard the current edge—based on whichever outcome does not increase the pessimistic estimator; such a choice must always exist for the pessimistic estimator to be valid. Once all edges are processed, the pessimistic estimator must still be a real number less than 11. But now, since there are no more undecided edges, the probability of failure is either 0 or 11. Since the pessimistic estimator is an upper bound which is less than 11, the probability of failure must be 0; in other words, the set of chosen edges is indeed a sparsifier of the graph.

In order for this de-randomization procedure to be efficient, the pessimistic estimator must be quickly evaluated and updated after considering each edge. Unfortunately, the probability union bound in the Benczur-Karger analysis involves all cuts in the graph, and is therefore an expression of exponential size and too expensive to serve as our pessimistic estimator. To design a more efficient pessimistic estimator, we need a more compact, easy-to-compute union bound over all cuts of the graph. We accomplish this by grouping all cuts of the graph into two types: small cuts and large cuts.

Small cuts.

Recall that our goal is to preserve cuts in the graph up to a (1+ϵ)(1+\epsilon) factor. Let us first restrict ourselves to all α\alpha-approximate mincuts of the graph for some α=no(1)\alpha=n^{o(1)}. There can be nΩ(α)n^{\Omega(\alpha)} many such cuts, so the naive union bound is still too slow. Here, our main strategy is to establish a structural representation of all α\alpha-approximate mincuts of a graph, with the goal of deriving a more compact “union bound” over all α\alpha-approximate cuts. This structure is built from an expander hierarchy of the graph, which is a hierarchical partitioning of the graph into disjoint expanders introduced by Goranci et al. [GRST20]. The connection between expanders and the mincut problem has been observed before [KT18, LP20]: in an expander with conductance ϕ\phi, all α\alpha-approximate mincuts must have at most α/ϕ\alpha/\phi vertices on one side, so a compact representation is simply all cuts with at most α/ϕ\alpha/\phi vertices on one side. Motivated by this connection, we show that if the original graph is itself an expander, then it is enough to preserve all vertex degrees and all edge weights up to an additive ϵλ\epsilon^{\prime}\lambda factor, where λ\lambda is the mincut of the graph and ϵ\epsilon^{\prime} depends on ϵ,α,ϕ\epsilon,\alpha,\phi. We present the unweighted expander case in Section 2 as a warm-up, which features all of our ideas except for the final expander decomposition step. To handle general graphs, we exploit the full machinery of the expander hierarchy [GRST20].

Large cuts.

For the large cuts—those that are not α\alpha-approximate mincuts—our strategy differs from the pessimistic estimator approach. Here, our aim is not to preserve each of them up to a (1+ϵ)(1+\epsilon)-factor, but a γ\gamma-factor for a different parameter γ=no(1)\gamma=n^{o(1)}. This relaxation prevents us from obtaining a full (1+ϵ)(1+\epsilon)-approximate sparsification of the graph, but it still works for the mincut problem as long as the large cuts do not fall below the original mincut value. While a deterministic (1+ϵ)(1+\epsilon)-approximate sparsification algorithm in near-linear time is unknown, one exists for γ\gamma-approximation sparsification for some γ=no(1)\gamma=n^{o(1)} [CGL+19]. In our case, we actually need the sparsifier to be uniformly weighted, so we construct our own sparsifier in Section 3.2.2, again via the expander hierarchy. Note that if the original graph is an expander, then we can take any expander whose degrees are roughly the same; in particular, the sparsifier does not need to be a subgraph of the original graph. To summarize, for the large cuts case, we simply construct an γ\gamma-approximate sparsifier deterministically, bypassing the need to de-randomize the Benczur-Karger random sampling technique.

Combining them together.

Of course, this γ\gamma-approximate sparsifier destroys the guarantee of the small cuts, which need to be preserved (1+ϵ)(1+\epsilon)-approximately. Our strategy is to combine the small cut sparsifier and the large cut sparsifier together in the following way. We take the union of the small cut sparsifier with a “lightly” weighted version of the large cut sparsifier, where each edge in it is weighted by ϵ/γ\epsilon/\gamma times its normal weight. This way, each small cut of weight ww suffers at most an additive γwϵ/γ=ϵw\gamma w\cdot\epsilon/\gamma=\epsilon w weight from the “light” large cut sparsifier, so we do not destroy the small cuts guarantee (up to replacing ϵ\epsilon with 2ϵ2\epsilon). Moreover, each large cut of weight wαλw\geq\alpha\lambda is weighted by at least w/γϵ/γαλ/γϵ/γ=α/γ2ϵλw/\gamma\cdot\epsilon/\gamma\geq\alpha\lambda/\gamma\cdot\epsilon/\gamma=\alpha/\gamma^{2}\cdot\epsilon\lambda, where λ\lambda is the mincut of the original graph. Hence, as long as αγ2/ϵ\alpha\geq\gamma^{2}/\epsilon, the large cuts have weight at least the mincut, and the property for large cuts is preserved.

Unbalanced vs. balanced.

We remark that our actual separation between small cuts and large cuts is somewhat different; we use unbalanced and balanced instead to emphasize this distinction. Nevertheless, we should intuitively think of unbalanced cuts as having small weight and balanced as having large weight; rather, the line is not drawn precisely at a weight threshold of αλ\alpha\lambda. The actual separation is more technical, so we omit it in this overview section.

1.2 Preliminaries

In this paper, all graphs are undirected, and nn and mm denote the number of vertices and edges of the input graph in question. All graphs are either unweighted or weighted multigraphs with polynomially bounded edge weights, i.e., in the range [1poly(n),poly(n)][\frac{1}{\textup{poly}(n)},\textup{poly}(n)]. We emphasize that even weighted graphs are multigraphs, which we find more convenient to work with.

We begin with more standard notation. For an unweighted graph G=(V,E)G=(V,E) and vertices u,vVu,v\in V, let #(u,v)\#(u,v) be the number of edges eEe\in E with endpoints uu and vv. For a weighted graph G=(V,E)G=(V,E) and edge eEe\in E, let w(e)w(e) be the weight of the edge, and for vertices u,vVu,v\in V, let w(u,v)w(u,v) be the sum of the weights w(e)w(e) of all (parallel) edges ee between uu and vv. For disjoint sets of vertices S,TVS,T\subseteq V, define E(S,T)EE(S,T)\subseteq E as the set of edges with one endpoint in SS and the other in TT, and define S:=E(S,VS)\partial S:=E(S,V\setminus S). For a set FEF\subseteq E of edges, denote its cardinality by |F||F| if GG is unweighted, and its total weight by w(F)w(F) if GG is weighted. Define the degree deg(v)\deg(v) of vertex vVv\in V to be |({v})||\partial(\{v\})| if GG is unweighted, and w(({v}))w(\partial(\{v\})) if GG is weighted. For a set SVS\subseteq V, define vol(S):=vSdeg(v)\textbf{{vol}}(S):=\sum_{v\in S}\deg(v). A cut of GG is the set of edges S\partial S for some SV\emptyset\subsetneq S\subsetneq V, and the mincut of GG is the cut S\partial S in GG that minimizes |S||\partial S| or w(S)w(\partial S) depending on if GG is unweighted or weighted. When the graph GG is ambiguous, we may add a subscript of GG in our notation, such as #G(u,v)\#_{G}(u,v).

1.2.1 Karger’s Approach

In this section, we outline Karger’s approach to his near-linear time randomized mincut algorithm and set up the necessary theorems for our deterministic result. Karger’s algorithm has two main steps. First, it computes a small set of (unweighted) trees on vertex set VV such that the mincut 22-respects one of the trees TT, defined as follows:

Definition 1.2.

Given a weighted graph GG and an unweighted tree TT on the same set of vertices, a cut GS\partial_{G}S 22-respects the tree TT if |TS|2|\partial_{T}S|\leq 2.

Karger accomplishes this goal by first sparsifying the graph into an unweighted skeleton graph using the well-known Benzcur-Karger sparsification by random sampling, and then running a tree packing algorithm of Gabow [Gab95] on the skeleton graph.

Theorem 1.3 (Karger [Kar00]).

Let GG be a weighted graph, let mm^{\prime} and cc^{\prime} be parameters, and let HH be an unweighted graph on the same vertices, called the skeleton graph, with the following properties:

  1. (a)

    HH has mm^{\prime} edges,

  2. (b)

    The mincut of HH is cc^{\prime}, and

  3. (c)

    The mincut in GG corresponds (under the same vertex partition) to a 7/6\nicefrac{{7}}{{6}}-approximate mincut in HH.

Given graphs GG and HH, there is a deterministic algorithm in O(cmlogn)O(c^{\prime}m^{\prime}\log n) time that constructs O(c)O(c^{\prime}) trees on the same vertices such that one of them 22-respects the mincut in GG.

The second main step of Karger’s algorithm is to compute the mincut of GG given a tree that 22-respects the mincut. This step is deterministic and is based on dynamic programming.

Theorem 1.4 (Karger [Kar00]).

Given a weighted, undirected graph GG and a (not necessarily spanning) tree TT on the same vertices, there is a deterministic algorithm in O(mlog2n)O(m\log^{2}n) time that computes the minimum-weight cut in GG that 22-respects the tree TT.

Our main technical contribution is a deterministic construction of the skeleton graph used in Theorem 1.3. Instead of designing an algorithm to produce the skeleton graph directly, it is more convenient to prove the following, which implies a skeleton graph by the following claim.

Theorem 1.5.

For any 0<ϵ10<\epsilon\leq 1, we can compute, in deterministic ϵ42O(logn)5/6(loglogn)O(1)m\epsilon^{-4}2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m time, an unweighted graph HH and some weight W=ϵ4λ/2O(logn)5/6(loglogn)O(1)W=\epsilon^{4}\lambda/2^{O(\log n)^{5/6}(\log\log n)^{O(1)}} such that

  1. 1.

    For any mincut S\partial S^{*} of GG, we have W|HS|(1+ϵ)λW\cdot|\partial_{H}S^{*}|\leq(1+\epsilon)\lambda, and

  2. 2.

    For any cut SV\emptyset\subsetneq S\subsetneq V of GG, we have W|HS|(1ϵ)λW\cdot|\partial_{H}S|\geq(1-\epsilon)\lambda.

Claim 1.6.

For ϵ=0.01\epsilon=0.01, the graph HH in Theorem 1.5 fulfills the conditions of Theorem 1.3 with m=m1+o(1)m^{\prime}=m^{1+o(1)} and c=no(1)c^{\prime}=n^{o(1)}.

Proof.

Since the algorithm of Theorem 1.5 takes m1+o(1)m^{1+o(1)} time, the output graph HH must have m1+o(1)m^{1+o(1)} edges, fulfilling condition (a) of Theorem 1.3. For any mincut SS^{*} of GG, by property (1) of Theorem 1.5, we have |HS|(1+ϵ)λ/Wno(1)|\partial_{H}S^{*}|\leq(1+\epsilon)\lambda/W\leq n^{o(1)}, fulfilling condition (b). For any cut SV\emptyset\subsetneq S\subsetneq V, by property (2), we have |HS|(1ϵ)λ/W|\partial_{H}S|\geq(1-\epsilon)\lambda/W. In other words, SS^{*} is a (1+ϵ)/(1ϵ)(1+\epsilon)/(1-\epsilon)-approximate mincut, which is a 7/6\nicefrac{{7}}{{6}}-approximate mincut for ϵ=0.01\epsilon=0.01, fulfilling condition (c). ∎

With the above three statements in hand, we now prove Theorem 1.1 following Karger’s approach. Run the algorithm of Theorem 1.5 to produce a graph HH which, by Claim 1.6, satisfies the conditions of Theorem 1.3. Apply Theorem 1.3 on GG and the skeleton graph HH, producing no(1)n^{o(1)} many trees such that one of them 22-respects the mincut in GG. Finally, run Theorem 1.4 on each tree separately and output the minimum 22-respecting cut found among all the trees, which must be the mincut in GG. Each step requires 2O(logn)5/6(loglogn)O(1)m2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m deterministic time, proving Theorem 1.1.

Thus, the main focus for the rest of the paper is proving Theorem 1.5.

1.2.2 Spectral Graph Theory

Central to our approach are the well-known concepts of conductance, expanders, and the graph Laplacian from spectral graph theory.

Definition 1.7 (Conductance, expander).

The conductance of a weighted graph GG is

Φ(G):=minSVw(E(S,VS))min{vol(S),vol(VS)}.\Phi(G):=\min_{\emptyset\subsetneq S\subsetneq V}\frac{w(E(S,V\setminus S))}{\min\{\textbf{{vol}}(S),\textbf{{vol}}(V\setminus S)\}}.

For the conductance of an unweighted graph, replace w(E(S,VS))w(E(S,V\setminus S)) by |E(S,VS)||E(S,V\setminus S)|. We say that GG is a ϕ\phi-expander if Φ(G)ϕ\Phi(G)\geq\phi.

Definition 1.8 (Laplacian).

The Laplacian LGL_{G} of a weighted graph G=(V,E)G=(V,E) is the n×nn\times n matrix, indexed by V×VV\times V, where

  1. (a)

    Each diagonal entry (v,v)(v,v) has entry deg(v)\deg(v), and

  2. (b)

    Each off-diagonal entry (u,v)(u,v) (uvu\neq v) has weight w(u,v)-w(u,v) if (u,v)E(u,v)\in E and 0 otherwise.

The only fact we will use about Laplacians is the following well-known fact, that cuts in graphs have the following nice form:

Fact 1.9.

For any weighted graph G=(V,E)G=(V,E) with Laplacian LGL_{G}, and for any subset SVS\subseteq V, we have

w(S)=𝟙STLG𝟙S,w(\partial S)=\mathbbm{1}_{S}^{T}L_{G}\mathbbm{1}_{S},

where 𝟙S{0,1}V\mathbbm{1}_{S}\in\{0,1\}^{V} is the vector with value 11 at vertex vv if vSv\in S, and value 0 otherwise. For unweighted graph GG, replace w(S)w(\partial S) with |S||\partial S|.

2 Expander Case

In this section, we prove Theorem 1.5 restricted to the case when GG is an unweighted expander. Our aim is to present an informal, intuitive exposition that highlights our main ideas in a relatively simple setting. Since this section is not technically required for the main result, we do not attempt to formalize our arguments, deferring the rigorous proofs to the general case in Section 3.

Theorem 2.1.

Let GG be an unweighted ϕ\phi-expander multigraph. For any 0<ϵ10<\epsilon\leq 1, we can compute, in deterministic m1+o(1)m^{1+o(1)} time, an unweighted graph HH and some weight W=ϵ3λ/no(1)W=\epsilon^{3}\lambda/n^{o(1)} such that

  1. (a)

    For any mincut GS\partial_{G}S^{*} of GG, we have W|HS|(1+ϵ)λW\cdot|\partial_{H}S^{*}|\leq(1+\epsilon)\lambda, and

  2. (b)

    For any cut GS\partial_{G}S of GG, we have W|HS|(1ϵ)λW\cdot|\partial_{H}S|\geq(1-\epsilon)\lambda.

For the rest of this section, we prove Theorem 2.1.

Consider an arbitrary cut GS\partial_{G}S. By Fact 1.9, we have

|GS|=𝟙STLG𝟙S=(vS𝟙vT)LG(vS𝟙v)=u,vS𝟙uTLG𝟙v.\displaystyle|\partial_{G}S|=\mathbbm{1}_{S}^{T}L_{G}\mathbbm{1}_{S}=\left(\sum_{v\in S}\mathbbm{1}_{v}^{T}\right)L_{G}\left(\sum_{v\in S}\mathbbm{1}_{v}\right)=\sum_{u,v\in S}\mathbbm{1}_{u}^{T}L_{G}\mathbbm{1}_{v}. (1)

Suppose we can approximate each 𝟙uTLG𝟙v\mathbbm{1}_{u}^{T}L_{G}\mathbbm{1}_{v} to an additive error of ϵλ\epsilon^{\prime}\lambda for some small ϵ\epsilon^{\prime} (depending on ϵ\epsilon); that is, suppose that our graph HH and weight WW satisfy

|𝟙uTLG𝟙vW𝟙uTLH𝟙v|ϵλ|\mathbbm{1}^{T}_{u}L_{G}\mathbbm{1}_{v}-W\cdot\mathbbm{1}_{u}^{T}L_{H}\mathbbm{1}_{v}|\leq\epsilon^{\prime}\lambda

for all u,vVu,v\in V. Then, by (1), we can approximate |GS||\partial_{G}S| up to an additive |S|2ϵλ|S|^{2}\epsilon^{\prime}\lambda, or a multiplicative (1+|S|2ϵ)(1+|S|^{2}\epsilon^{\prime}), which is good if |S||S| is small. Similarly, if |VS||V\setminus S| is small, then we can replace SS with VSV\setminus S in (1) and approximate |GS|=|G(VS)||\partial_{G}S|=|\partial_{G}(V\setminus S)| to the same factor. Motivated by this observation, we define a set SVS\subseteq V to be unbalanced if min{vol(S),vol(VS)}αλ/ϕ\min\{\textbf{{vol}}(S),\textbf{{vol}}(V\setminus S)\}\leq\alpha\lambda/\phi for some α=no(1)\alpha=n^{o(1)} to be set later. Similarly, define a cut GS\partial_{G}S to be unbalanced if the set SS is unbalanced. Note that an unbalanced set SS must have either |S|α/ϕ|S|\leq\alpha/\phi or |VS|α/ϕ|V\setminus S|\leq\alpha/\phi, since if we assume without loss of generality that vol(S)vol(VS)\textbf{{vol}}(S)\leq\textbf{{vol}}(V\setminus S), then

|S|λvSdeg(v)=vol(S)αλ/ϕ,\displaystyle|S|\lambda\leq\sum_{v\in S}\deg(v)=\textbf{{vol}}(S)\leq\alpha\lambda/\phi, (2)

where the first inequality uses that each degree cut ({v})\partial(\{v\}) has weight deg(v)λ\deg(v)\geq\lambda. Moreover, since GG is a ϕ\phi-expander, the mincut GS\partial_{G}S^{*} is unbalanced because, assuming without loss of generality that vol(S)vol(VS)\textbf{{vol}}(S^{*})\leq\textbf{{vol}}(V\setminus S^{*}), we obtain

|G(S)|vol(S)Φ(G)ϕvol(S)1/ϕαλ/ϕ.\frac{|\partial_{G}(S^{*})|}{\textbf{{vol}}(S^{*})}\geq\Phi(G)\geq\phi\implies\textbf{{vol}}(S^{*})\leq 1/\phi\leq\alpha\lambda/\phi.

To approximate all unbalanced cuts, it suffices by (1) and (2) to approximate each 𝟙uTLG𝟙v\mathbbm{1}^{T}_{u}L_{G}\mathbbm{1}_{v} up to additive error (ϕ/α)2ϵλ(\phi/\alpha)^{2}\epsilon\lambda. When uvu\neq v, the expression 𝟙uTLG𝟙v\mathbbm{1}_{u}^{T}L_{G}\mathbbm{1}_{v} is simply the negative of the number of parallel (u,v)(u,v) edges in GG. So, approximating 𝟙uTLG𝟙v\mathbbm{1}_{u}^{T}L_{G}\mathbbm{1}_{v} up to additive error ϵλ\epsilon\lambda simply amounts to approximating the number of parallel (u,v)(u,v) edges. When u=vu=v, the expression 𝟙vTLG𝟙v\mathbbm{1}_{v}^{T}L_{G}\mathbbm{1}_{v} is simply the degree of vv, so approximating it amounts to approximating the degree of vv.

Consider what happens if we randomly sample each edge with probability p=Θ(αlognϵ2ϕλ)p=\Theta(\frac{\alpha\log n}{\epsilon^{2}\phi\lambda}) and weight the sampled edges by W^:=1/p\widehat{W}:=1/p to form the sampled graph H^\widehat{H}. For the terms 𝟙uTLG𝟙v\mathbbm{1}_{u}^{T}L_{G}\mathbbm{1}_{v} (uvu\neq v), we have #G(u,v)vol(S)αλ/ϕ\#_{G}(u,v)\leq\textbf{{vol}}(S)\leq\alpha\lambda/\phi. Let us assume for simplicity that #G(u,v)=αλ/ϕ\#_{G}(u,v)=\alpha\lambda/\phi, which turns out to be the worst case. By Chernoff bounds, for δ=ϵϕ/α\delta=\epsilon\phi/\alpha,

Pr[|#H^(u,v)p#G(u,v)|>δp#G(u,v)]\displaystyle\Pr\left[\left|\#_{\widehat{H}}(u,v)-p\cdot\#_{G}(u,v)\right|>\delta\cdot p\cdot\#_{G}(u,v)\right] <2exp(δ2p#G(u,v)/3)\displaystyle<2\exp(-\delta^{2}\cdot p\cdot\#_{G}(u,v)/3)
=2exp((ϵϕα)2Θ(αlognϵ2ϕλ)αλ/ϕ3)\displaystyle=2\exp\left(-\left(\frac{\epsilon\phi}{\alpha}\right)^{2}\cdot\Theta\left(\frac{\alpha\log n}{\epsilon^{2}\phi\lambda}\right)\cdot\frac{\alpha\lambda/\phi}{3}\right) (3)
=2exp(Θ(logn)),\displaystyle=2\exp(-\Theta(\log n)),

which we can set to be much less than 1/n21/n^{2}. We then have the implication

|#H^(u,v)p#G(u,v)|δp#G(u,v)|𝟙uT(LGLH^)𝟙v|δ#G(u,v)=ϵϕ/ααλ/ϕ=ϵλ.\left|\#_{\widehat{H}}(u,v)-p\cdot\#_{G}(u,v)\right|\leq\delta\cdot p\cdot\#_{G}(u,v)\implies\left|\mathbbm{1}_{u}^{T}(L_{G}-L_{\widehat{H}})\mathbbm{1}_{v}\right|\leq\delta\cdot\#_{G}(u,v)=\epsilon\phi/\alpha\cdot\alpha\lambda/\phi=\epsilon\lambda.

Similarly, for the terms 𝟙vTLG𝟙v\mathbbm{1}_{v}^{T}L_{G}\mathbbm{1}_{v}, we have deg(v)vol(S)αλ/ϕ\deg(v)\leq\textbf{{vol}}(S)\leq\alpha\lambda/\phi, and the same calculation can be made.

From this random sampling analysis, we can derive the following pessimistic estimator. Initially, it is the sum of the quantities (3) for all (u,v)(u,v) satisfying either u=vu=v or (u,v)E(u,v)\in E. This sum has O(m)O(m) terms which sum to less than 11, so it can be efficiently computed and satisfies the initial condition of a pessimistic estimator. After some edges have been considered, the probability upper bounds (3) are modified to be conditional to the choices of edges so far, which can still be efficiently computed. At the end, for each unbalanced set SS, the graph H^\widehat{H} will satisfy

||GS|W^|H^S||ϵλ(1ϵ)|GS|W^|H^S|(1+ϵ)|GS|.\big{|}|\partial_{G}S|-\widehat{W}\cdot|\partial_{\widehat{H}}S|\big{|}\leq\epsilon\lambda\implies(1-\epsilon)|\partial_{G}S|\leq\widehat{W}\cdot|\partial_{\widehat{H}}S|\leq(1+\epsilon)|\partial_{G}S|.

Since any mincut GS\partial_{G}S^{*} is unbalanced, we fulfill condition (a) of Theorem 2.1. We also fulfill condition (b) for any cut with a side that is unbalanced. This concludes the unbalanced case; we omit the rest of the details, deferring the pessimistic estimator and its efficient computation to the general case, specifically Section 3.2.1.

Define a cut to be balanced if it is not unbalanced. For the balanced cuts, it remains to fulfill condition (b), which may not hold for the graph H^\widehat{H}. Our solution is to “overlay” a fixed expander onto the graph H^\widehat{H}, weighted small enough to barely affect the mincut (in order to preserve condition (a)), but large enough to force all balanced cuts to have weight at least λ\lambda. In particular, let H~\widetilde{H} be an unweighted Θ(1)\Theta(1)-expander on the same vertex set VV where each vertex vVv\in V has degree Θ(degG(v)/λ)\Theta(\deg_{G}(v)/\lambda), and let W~:=Θ(ϵϕλ)\widetilde{W}:=\Theta(\epsilon\phi\lambda). We should think of H~\widetilde{H} as a “lossy” sparsifier of GG, in that it approximates cuts up to factor O(1/ϕ)O(1/\phi), not (1+ϵ)(1+\epsilon).

Consider taking the “union” of the graph H^\widehat{H} weighted by W^\widehat{W} and the graph H~\widetilde{H} weighted by W~\widetilde{W}. More formally, consider a weighted graph HH^{\prime} where each edge (u,v)(u,v) is weighted by W^wH^(u,v)+W~wH~(u,v)\widehat{W}\cdot w_{\widehat{H}}(u,v)+\widetilde{W}\cdot w_{\widetilde{H}}(u,v). We now show two properties: (1) the mincut gains relatively little weight from H~\widetilde{H} in the union HH^{\prime}, and (2) any balanced cut automatically has at least λ\lambda total weight from H~\widetilde{H}.

  1. 1.

    For a mincut GS\partial_{G}S^{*} in GG with volG(S)|GS|/ϕ=λ/ϕ\textbf{{vol}}_{G}(S^{*})\leq|\partial_{G}S^{*}|/\phi=\lambda/\phi, the cut crosses

    w(H^S)volH^(S)Θ(1)volG(S)/λΘ(1/ϕ)w(\partial_{\widehat{H}}S^{*})\leq\textbf{{vol}}_{\widehat{H}}(S^{*})\leq\Theta(1)\cdot\textbf{{vol}}_{G}(S^{*})/\lambda\leq\Theta(1/\phi)

    edges in H~\widetilde{H}, for a total cost of at most Θ(1/ϕ)Θ(ϵϕλ)ϵλ\Theta(1/\phi)\cdot\Theta(\epsilon\phi\lambda)\leq\epsilon\lambda.

  2. 2.

    For a balanced cut GS\partial_{G}S, it satisfies |GS|ϕvolG(S)αλ|\partial_{G}S|\geq\phi\cdot\textbf{{vol}}_{G}(S)\geq\alpha\lambda, so it crosses

    w(H^S)Θ(1)volH^(S)Θ(1)volG(S)/λΘ(α/ϕ)w(\partial_{\widehat{H}}S)\geq\Theta(1)\cdot\textbf{{vol}}_{\widehat{H}}(S)\geq\Theta(1)\cdot\textbf{{vol}}_{G}(S)/\lambda\geq\Theta(\alpha/\phi)

    many edges in H~\widetilde{H}, for a total cost of at least Θ(α/ϕ)Θ(ϵϕλ)\Theta(\alpha/\phi)\cdot\Theta(\epsilon\phi\lambda). Setting α:=Θ(1ϵ)\alpha:=\Theta(\frac{1}{\epsilon}), the cost becomes at least λ\lambda.

Therefore, in the weighted graph HH^{\prime}, the mincut has weight at most (1+O(ϵ))λ(1+O(\epsilon))\lambda, and any cut has weight at least (1ϵ)λ(1-\epsilon)\lambda. We can reset ϵ\epsilon to be a constant factor smaller so that the factor (1+O(ϵ))(1+O(\epsilon)) becomes (1+ϵ)(1+\epsilon).

To finish the proof of Theorem 2.1, it remains to extract an unweighted graph HH and a weight WW from the weighted graph HH^{\prime}. Since W^=Θ(ϵ2ϕλαlogn)=Θ(ϵ3ϕλlogn)\widehat{W}=\Theta(\frac{\epsilon^{2}\phi\lambda}{\alpha\log n})=\Theta(\frac{\epsilon^{3}\phi\lambda}{\log n}) and W~=Θ(ϵϕλ)\widetilde{W}=\Theta(\epsilon\phi\lambda), we can make W~\widetilde{W} an integer multiple of W^\widehat{W}, so that each edge in HH^{\prime} is an integer multiple of W^\widehat{W}. We can therefore set W:=W^W:=\widehat{W} and define the unweighted graph HH so that #H(u,v)=wH(u,v)/W^\#_{H}(u,v)=w_{H^{\prime}}(u,v)/\widehat{W} for all u,vVu,v\in V.

3 General Case

This section is dedicated to proving Theorem 1.5. For simplicity, we instead prove the following restricted version first, which has the additional assumption that the maximum edge weight in GG is bounded. At the end of this section, we show why this assumption can be removed to obtain the full Theorem 1.5.

Theorem 3.1.

There exists a function f(n)2O(logn)5/6(loglogn)O(1)f(n)\leq 2^{O(\log n)^{5/6}(\log\log n)^{O(1)}} such that the following holds. Let GG be a graph with mincut λ\lambda and maximum edge weight at most ϵ4λ/f(n)\epsilon^{4}\lambda/f(n). For any 0<ϵ10<\epsilon\leq 1, we can compute, in deterministic 2O(logn)5/6(loglogn)O(1)m2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m time, an unweighted graph HH and some weight Wϵ4λ/f(n)W\geq\epsilon^{4}\lambda/f(n) such that the two properties of Theorem 1.5 hold, i.e.,

  1. 1.

    For any mincut SS^{*} of GG, we have W|HS|(1+ϵ)λW\cdot|\partial_{H}S^{*}|\leq(1+\epsilon)\lambda, and

  2. 2.

    For any cut SV\emptyset\subsetneq S\subsetneq V of GG, we have W|HS|(1ϵ)λW\cdot|\partial_{H}S|\geq(1-\epsilon)\lambda.

3.1 Expander Decomposition Preliminaries

Our main tool in generalizing the expander case is expander decompositions, which was popularized by Spielman and Teng [ST04] and is quickly gaining traction in the area of fast graph algorithms. The general approach to utilizing expander decompositions is as follows. First, solve the case when the input graph is an expander, which we have done in Section 2 for the problem described in Theorem 1.5. Then, for a general graph, decompose it into a collection of expanders with few edges between the expanders, solve the problem each expander separately, and combine the solutions together, which often involves a recursive call on a graph that is a constant-factor smaller. For our purposes, we use a slightly stronger variant than the usual expander decomposition that ensures boundary-linkedness, which will be important in our analysis. The following definition is inspired by [GRST20]; note that our variant is weaker than the one in Definition 4.2 of [GRST20] in that we only guarantee their property (2). For completeness, we include a full proof in Appendix A that is similar to the one in [GRST20], and assuming a subroutine called WeightedBalCutPrune from [LS21].

Theorem 3.2 (Boundary-linked expander decomposition).

Let G=(V,E)G=(V,E) be a graph and let r1r\geq 1 be a parameter. There is a deterministic algorithm in m1+O(1/r)+O~(m/ϕ2)m^{1+O(1/r)}+\widetilde{O}(m/\phi^{2}) time that, for any parameters β(logn)O(r4)\beta\leq(\log n)^{-O(r^{4})} and ϕβ\phi\leq\beta, partitions V=V1VkV=V_{1}\uplus\cdots\uplus V_{k} such that

  1. 1.

    Each vertex set ViV_{i} satisfies

    minSViw(G[Vi]S)min{volG[Vi](S)+βϕw(EG(S,VVi)),volG[Vi](ViS)+βϕw(EG(ViS,VVi))}ϕ.\displaystyle\min_{\emptyset\subsetneq S\subsetneq V_{i}}\frac{w(\partial_{G[V_{i}]}S)}{\min\{\textbf{{vol}}_{G[V_{i}]}(S)+\frac{\beta}{\phi}w(E_{G}(S,V\setminus V_{i})),\textbf{{vol}}_{G[V_{i}]}(V_{i}\setminus S)+\frac{\beta}{\phi}w(E_{G}(V_{i}\setminus S,V\setminus V_{i}))\}}\geq\phi. (4)

    Informally, we call the graph G[Vi]G[V_{i}] together with its boundary edges EG(Vi,VVi)E_{G}(V_{i},V\setminus V_{i}) a β\beta-boundary-linked ϕ\phi-expander.222For unweighted graphs, [GRST20] uses the notation G[Vi]β/ϕG[V_{i}]^{\beta/\phi} to represent a graph where each (boundary) edge in E(Vi,VVi)E(V_{i},V\setminus V_{i}) is replaced with β/ϕ\beta/\phi many self-loops at the endpoint in ViV_{i}. With this definition, (4) is equivalent to saying that G[Vi]β/ϕG[V_{i}]^{\beta/\phi} is a ϕ\phi-expander. We will use this definition when proving Theorem 3.2 in Appendix A. In particular, for any SS satisfying

    volG[Vi](S)+βϕw(EG(S,VVi))volG[Vi](ViS)+βϕw(EG(ViS,VVi)),\textbf{{vol}}_{G[V_{i}]}(S)+\frac{\beta}{\phi}w(E_{G}(S,V\setminus V_{i}))\leq\textbf{{vol}}_{G[V_{i}]}(V_{i}\setminus S)+\frac{\beta}{\phi}w(E_{G}(V_{i}\setminus S,V\setminus V_{i})),

    we simultaneously obtain

    w(G[Vi]S)volG[Vi](S)ϕandw(G[Vi]S)βϕw(EG(S,VVi))ϕw(G[Vi]S)w(EG(S,VVi))β.\frac{w(\partial_{G[V_{i}]}S)}{\textbf{{vol}}_{G[V_{i}]}(S)}\geq\phi\qquad\text{and}\qquad\frac{w(\partial_{G[V_{i}]}S)}{\frac{\beta}{\phi}w(E_{G}(S,V\setminus V_{i}))}\geq\phi\iff\frac{w(\partial_{G[V_{i}]}S)}{w(E_{G}(S,V\setminus V_{i}))}\geq\beta.

    The right-most inequality is where the name “boundary-linked” comes from.

  2. 2.

    The total weight of “inter-cluster” edges, w(V1Vk)w(\partial V_{1}\cup\cdots\cup\partial V_{k}), is at most (logn)O(r4)ϕvol(V)(\log n)^{O(r^{4})}\phi\textbf{{vol}}(V).

Note that for our applications, it’s important that the boundary-linked parameter β\beta is much larger than ϕ\phi. This is because in our recursive algorithm, the approximation factor will blow up by roughly 1/β1/\beta per recursion level, while the instance size shrinks by roughly ϕ\phi.

In order to capture recursion via expander decompositions, we now define a boundary-linked expander decomposition sequence {Gi}\{G^{i}\} on the graph GG in a similar way to [GRST20]. Compute a boundary-linked expander decomposition for β\beta and ϕβ\phi\leq\beta to be determined later, contract each expander,333Since we are working with weighted multigraphs, we do not collapse parallel edges obtained from contraction into single edges. and recursively decompose the contracted graph until the graph consists of a single vertex. Let G0=GG^{0}=G be the original graph and G1,G2,,GLG^{1},G^{2},\ldots,G^{L} be the recursive contracted graphs. Note that each graph GiG^{i} has minimum degree at least λ\lambda, since any degree cut in any GiG^{i} induces a cut in the original graph GG. Each time we contract, we will keep edge identities for the edges that survive, so that E(G0)E(G1)E(GL)E(G^{0})\supseteq E(G^{1})\supseteq\cdots\supseteq E(G^{L}). Let UiU^{i} be the vertices of GiG^{i}.

For the rest of Section 3.1, fix an expander decomposition sequence {Gi}\{G^{i}\} of GG. For any subset SV\emptyset\subsetneq S\subsetneq V, we now define an decomposition sequence of SS as follows. Let S0=SS^{0}=S, and for each i>0i>0, construct Si+1S^{i+1} as a subset of the vertices of Gi+1G^{i+1}, as follows. Take the expander decomposition of GiG^{i}, which partitions the vertices UiU^{i} of GiG^{i} into, say, U1i,,UkiiU^{i}_{1},\ldots,U^{i}_{k_{i}}. Each of the UjiU^{i}_{j} gets contracted to a single vertex uju_{j} in GiG^{i}. For each UjiU^{i}_{j}, we have a choice whether to add uju_{j} to SiS^{i} or not. This completes the construction of SiS^{i}. Define the “difference” Dji=UjSiD^{i}_{j}=U_{j}\setminus S^{i} if ujSiu_{j}\in S^{i}, and Dji=UjSiD^{i}_{j}=U_{j}\cap S^{i} otherwise. The sets SiS^{i}, UjiU^{i}_{j}, and DjiD^{i}_{j} define the decomposition sequence of SS.

We now prove some key properties of the boundary-linked expander decomposition sequence in the context of graph cuts, which we will use later on. First, regardless of the choice whether to add each uju_{j} to SiS^{i}, we have the following lemma relating the sets DjiD^{i}_{j} to the original set SS.

Lemma 3.3.

For any decomposition sequence {Si}\{S^{i}\} of SS,

GSi=0Lj[ki]GiDji.\partial_{G}S\subseteq\bigcup_{i=0}^{L}\bigcup_{j\in[k_{i}]}\partial_{G^{i}}D^{i}_{j}.
Proof.

Observe that

(GiSi)(Gi+1Si+1)j[ki]GiDji.\displaystyle(\partial_{G^{i}}S^{i})\triangle(\partial_{G^{i+1}}S^{i+1})\subseteq\bigcup_{j\in[k_{i}]}\partial_{G^{i}}D^{i}_{j}. (5)

In particular,

GiSiGi+1Si+1j[ki]GiDji.\partial_{G^{i}}S^{i}\subseteq\partial_{G^{i+1}}S^{i+1}\cup\bigcup_{j\in[k_{i}]}\partial_{G^{i}}D^{i}_{j}.

Iterating this over all ii,

GSi=0Lj[ki]GiDji.\partial_{G}S\subseteq\bigcup_{i=0}^{L}\bigcup_{j\in[k_{i}]}\partial_{G^{i}}D^{i}_{j}.

We now define a specific decomposition sequence of SS, by setting up the rule whether or not to include each uju_{j} in SiS^{i}. For each UjiU^{i}_{j}, if

volGi[Uji](SiUji)+βϕw(EGi(SiUji,UiUji))volGi[Uji](UjiSi)+βϕw(EGi(UjiSi,UiUji)),\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(S^{i}\cap U^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(S^{i}\cap U^{i}_{j},U^{i}\setminus U^{i}_{j}))\geq\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(U^{i}_{j}\setminus S^{i})+\frac{\beta}{\phi}w(E_{G^{i}}(U^{i}_{j}\setminus S^{i},U^{i}\setminus U^{i}_{j})),

then add uju_{j} to SiS^{i}; otherwise, do not add uju_{j} to SiS^{i}. This ensures that

volGi[Uji](UjiDji)+βϕw(EGi(UjiDji,UiUji))volGi[Uji](Dji)+βϕw(EGi(Dji,UiUji)).\displaystyle\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(U^{i}_{j}\setminus D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j},U^{i}\setminus U^{i}_{j}))\geq\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})). (6)

Since Gi[Uji]G^{i}[U^{i}_{j}] is a β\beta-boundary-linked ϕ\phi-expander, by our construction, we have, for all i,ji,j,

w(Gi[Uji]Dji)volGi[Uji](Dji)ϕ\displaystyle\frac{w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})}{\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})}\geq\phi (7)

and

w(Gi[Uji]Dji)w(EGi(Dji,UiUji))β.\displaystyle\frac{w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})}{w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))}\geq\beta. (8)

For this specific construction of {Si}\{S^{i}\}, called the canonical decomposition sequence of SS, we have the following lemma, which complements Lemma 3.3.

Lemma 3.4.

Let {Si}\{S^{i}\} be any decomposition sequence of SS satisfying (8) for all i,ji,j. Then,

i=0Lj[ki]w(GiDji)βO(L)w(GS).\sum_{i=0}^{L}\sum_{j\in[k_{i}]}w(\partial_{G^{i}}D^{i}_{j})\leq\beta^{-O(L)}w(\partial_{G}S).
Proof.

By (8),

w(EGi(Dji,UiUji))1βw(Gi[Uji]Dji).w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\leq\frac{1}{\beta}\cdot w(\partial_{G^{i}[U^{i}_{j}]}D_{j}^{i}).

The edges of Gi[Uji]Dji\partial_{G^{i}[U^{i}_{j}]}D_{j}^{i} are inside GiSi\partial_{G^{i}}S^{i} and are disjoint over distinct jj, so in total,

j[ki]w(GiDji)j[ki]1βw(Gi[Uji]Dji)1βw(GiSi).\sum_{j\in[k_{i}]}w(\partial_{G^{i}}D^{i}_{j})\leq\sum_{j\in[k_{i}]}\frac{1}{\beta}\cdot w(\partial_{G^{i}[U^{i}_{j}]}D_{j}^{i})\leq\frac{1}{\beta}\cdot w(\partial_{G^{i}}S^{i}).

From (5), we also obtain

Gi+1Si+1GiSij[ki]GiDji.\partial_{G^{i+1}}S^{i+1}\subseteq\partial_{G^{i}}S^{i}\cup\bigcup_{j\in[k_{i}]}\partial_{G^{i}}D^{i}_{j}.

Therefore,

w(Gi+1Si+1)w(GiSi)+w(j[ki]GiDji)(1+1β)w(GiSi).w(\partial_{G^{i+1}}S^{i+1})\leq w(\partial_{G^{i}}S^{i})+w\left(\bigcup_{j\in[k_{i}]}\partial_{G^{i}}D^{i}_{j}\right)\leq\left(1+\frac{1}{\beta}\right)\cdot w(\partial_{G^{i}}S^{i}).

Iterating this over all i[L]i\in[L], we obtain

w(GiSi)(1+1β)iw(GS).w(\partial_{G^{i}}S^{i})\leq\left(1+\frac{1}{\beta}\right)^{i}\cdot w(\partial_{G}S).

Thus,

i=0Lj[ki]w(GiDji)i=0L1βw(GiSi)i=0L1β(1+1β)iw(GS)=βO(L)w(GS).\sum_{i=0}^{L}\sum_{j\in[k_{i}]}w(\partial_{G^{i}}D^{i}_{j})\leq\sum_{i=0}^{L}\frac{1}{\beta}\cdot w(\partial_{G^{i}}S^{i})\leq\sum_{i=0}^{L}\frac{1}{\beta}\cdot\left(1+\frac{1}{\beta}\right)^{i}\cdot w(\partial_{G}S)=\beta^{-O(L)}w(\partial_{G}S).

3.2 Unbalanced Case

In this section, we generalize the notion of unbalanced from Section 2 to the general case, and then prove a (1+ϵ)(1+\epsilon)-approximate sparsifier of the unbalanced cuts.

Fix an expander decomposition sequence {Gi}\{G^{i}\} of GG for the Section 3.2. For a given set SV\emptyset\subsetneq S\subsetneq V, let {Si}\{S^{i}\} be the canonical decomposition sequence of SS, and define DjiD^{i}_{j} as before, so that they satisfy (7) and (8) for all i,ji,j. We generalize our definition of unbalanced from the expander case as follows, for some τ=no(1)\tau=n^{o(1)} to be specified later.

Definition 3.5.

The set SVS\subseteq V is τ\tau-unbalanced if for each level ii, j[ki]volGi(Dji)τλ/ϕ\sum_{j\in[k_{i}]}\textbf{{vol}}_{G^{i}}(D^{i}_{j})\leq\tau\lambda/\phi. A cut S\partial S is τ\tau-unbalanced if the set SS is τ\tau-unbalanced.

Note that if GG is originally an expander, then in the first expander decomposition of the sequence, we can declare the entire graph as a single expander; in this case, the expander decomposition sequence stops immediately, and the definition of τ\tau-unbalanced becomes equivalent to that from the expander case. We now claim that for an appropriate value of τ\tau, any mincut is τ\tau-unbalanced.

Claim 3.6.

For τβΩ(L)\tau\geq\beta^{-\Omega(L)}, any mincut S\partial S^{*} of GG is τ\tau-unbalanced.

Proof.

Consider the canonical decomposition sequence of SS, and define DjiD^{i}_{j} as usual. For each level ii and index j[ki]j\in[k_{i}],

volGi(Dji)\displaystyle\textbf{{vol}}_{G^{i}}(D^{i}_{j}) =volGi[Uji](Dji)+w(EGi(Dji,UiUji))\displaystyle=\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
(7)1ϕw(Gi[Uji]Dji)+w(EGi(Dji,UiUji))\displaystyle\stackrel{{\scriptstyle\mathclap{(\ref{eq:Exp})}}}{{\leq}}\frac{1}{\phi}w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
1ϕw(GiDji).\displaystyle\leq\frac{1}{\phi}w(\partial_{G^{i}}D^{i}_{j}).

Summing over all j[ki]j\in[k_{i}] and applying Lemma 3.4,

j[ki]volGi(Dji)j[ki]1ϕw(GiDji)=1ϕj[ki]w(GiDji)Lem.3.41ϕβO(L)w(GS)τλϕ,\sum_{j\in[k_{i}]}\textbf{{vol}}_{G^{i}}(D^{i}_{j})\leq\sum_{j\in[k_{i}]}\frac{1}{\phi}w(\partial_{G^{i}}D^{i}_{j})=\frac{1}{\phi}\cdot\sum_{j\in[k_{i}]}w(\partial_{G^{i}}D^{i}_{j})\stackrel{{\scriptstyle\textup{Lem.}\ref{lem:D-ub}}}{{\leq}}\frac{1}{\phi}\cdot\beta^{-O(L)}w(\partial_{G}S^{*})\leq\frac{\tau\lambda}{\phi},

so SS^{*} is τ\tau-unbalanced. ∎

Let us now introduce some notation exclusive to this section. For each vertex vUiv\in U^{i}, let v¯V\overline{v}\subseteq V be its “pullback” on the original set VV, defined as all vertices in VV that get contracted into vv in graph GiG^{i} in the expander sequence. For each set DjiD^{i}_{j}, let Dji¯V\overline{D^{i}_{j}}\subseteq V be the pullback of DjiD^{i}_{j}, defined as Dji¯=vDjiv¯\overline{D^{i}_{j}}=\bigcup_{v\in D^{i}_{j}}\overline{v}. We can then write

𝟙S=i,j±𝟙Dji¯=i,jvDji±𝟙v¯,\mathbbm{1}_{S}=\sum_{i,j}\pm\mathbbm{1}_{\overline{D^{i}_{j}}}=\sum_{i,j}\sum_{v\in D^{i}_{j}}\pm\mathbbm{1}_{\overline{v}},

where the ±\pm sign depends on whether Dji=UjiSiD^{i}_{j}=U^{i}_{j}\setminus S^{i} or Dji=UjiSiD^{i}_{j}=U^{i}_{j}\cap S^{i}. Then,

w(GS)=𝟙STLG𝟙S=i,j,k,l±𝟙Dji¯TLG𝟙Dlk¯=i,j,k,luDji,vDlk±𝟙u¯TLG𝟙v¯.\displaystyle w(\partial_{G}S)=\mathbbm{1}_{S}^{T}L_{G}\mathbbm{1}_{S}=\sum_{i,j,k,l}\pm\mathbbm{1}_{\overline{D^{i}_{j}}}^{T}L_{G}\mathbbm{1}_{\overline{D^{k}_{l}}}=\sum_{i,j,k,l}\sum_{u\in D^{i}_{j},v\in D^{k}_{l}}\pm\mathbbm{1}^{T}_{\overline{u}}L_{G}\mathbbm{1}_{\overline{v}}. (9)
Claim 3.7.

For an τ\tau-unbalanced set SS, there are at most ((L+1)τ/ϕ)2((L+1)\tau/\phi)^{2} nonzero terms in the summation (9).

Proof.

Each vertex vDjiv\in D^{i}_{j} has degree at least λ\lambda in GiG^{i}, since it induces a cut (specifically, its pullback v¯V\overline{v}\subseteq V) in the original graph GG. Therefore,

τλ/ϕj[ki]volGi(Dji)j[ki]|Dji|λ,\tau\lambda/\phi\geq\sum_{j\in[k_{i}]}\textbf{{vol}}_{G^{i}}(D^{i}_{j})\geq\sum_{j\in[k_{i}]}|D^{i}_{j}|\cdot\lambda,

so there are at most τ/ϕ\tau/\phi many choices for jj and uDjiu\in D^{i}_{j} given a level ii. There are at most L+1L+1 many choices for ii, giving at most (L+1)τ/ϕ(L+1)\tau/\phi many combinations of i,j,ui,j,u. The same holds for combinations of k,l,vk,l,v, hence the claim. ∎

The main goal of this section is to prove the following lemma.

Lemma 3.8.

There exists a constant C>0C>0 such that given any weight WCϵϕλτln(Lm)W\leq\frac{C\epsilon\phi\lambda}{\tau\ln(Lm)}, we can compute, in deterministic O~(L2m)\widetilde{O}(L^{2}m) time,444outside of computing the boundary-linked expander decomposition sequence an unweighted graph HH such that for all levels i,ki,k and vertices uUi,vUku\in U^{i},v\in U^{k} satisfying degGi(u)τλ/ϕ\deg_{G^{i}}(u)\leq\tau\lambda/\phi and degGk(v)τλ/ϕ\deg_{G^{k}}(v)\leq\tau\lambda/\phi,

|𝟙u¯TLG𝟙v¯W𝟙u¯TLH𝟙v¯|ϵλ.\displaystyle\left|\mathbbm{1}^{T}_{\overline{u}}L_{G}\mathbbm{1}_{\overline{v}}-W\cdot\mathbbm{1}^{T}_{\overline{u}}L_{H}\mathbbm{1}_{\overline{v}}\right|\leq\epsilon\lambda. (10)

Before we prove Lemma 3.8, we show that it implies a sparsifier of τ\tau-unbalanced cuts, which is the lemma we will eventually use to prove Theorem 3.1:

Lemma 3.9.

There exists a constant C>0C>0 such that given any weight WCϵϕλτln(Lm)W\leq\frac{C\epsilon\phi\lambda}{\tau\ln(Lm)}, we can compute, in deterministic O~(L2m)\widetilde{O}(L^{2}m) time, an unweighted graph HH such that for each τ\tau-unbalanced cut SS,

|w(GS)Ww(HS)|((L+1)τϕ)2ϵλ.\big{|}w(\partial_{G}S)-W\cdot w(\partial_{H}S)\big{|}\leq\left(\frac{(L+1)\tau}{\phi}\right)^{2}\cdot\epsilon\lambda.
Proof.

Let C>0C>0 be the same constant as the one in Lemma 3.8. Applying (9) to HS\partial_{H}S as well, we have

w(GS)Ww(HS)=i,j,k,luDji,vDlk±(𝟙u¯TLG𝟙v¯W𝟙u¯TLH𝟙v¯),w(\partial_{G}S)-W\cdot w(\partial_{H}S)=\sum_{i,j,k,l}\sum_{u\in D^{i}_{j},v\in D^{k}_{l}}\pm(\mathbbm{1}^{T}_{\overline{u}}L_{G}\mathbbm{1}_{\overline{v}}-W\cdot\mathbbm{1}^{T}_{\overline{u}}L_{H}\mathbbm{1}_{\overline{v}}),

so that

|w(GS)Ww(HS)|i,j,k,luDji,vDlk|𝟙u¯TLG𝟙v¯W𝟙u¯TLH𝟙v¯|.\big{|}w(\partial_{G}S)-W\cdot w(\partial_{H}S)\big{|}\leq\sum_{i,j,k,l}\sum_{u\in D^{i}_{j},v\in D^{k}_{l}}\big{|}\mathbbm{1}^{T}_{\overline{u}}L_{G}\mathbbm{1}_{\overline{v}}-W\cdot\mathbbm{1}^{T}_{\overline{u}}L_{H}\mathbbm{1}_{\overline{v}}\big{|}.

By Claim 3.7, there are at most ((L+1)τ/ϕ)2((L+1)\tau/\phi)^{2} nonzero terms in the summation above. In order to apply Lemma 3.8 to each such term, we need to show that degGi(u)τλ/ϕ\deg_{G^{i}}(u)\leq\tau\lambda/\phi and degGk(v)τλ/ϕ\deg_{G^{k}}(v)\leq\tau\lambda/\phi. Since SS is an τ\tau-unbalanced cut, we have

degGi(u)volGi(Dji)j[ki]volGi(Dji)τλ/ϕ,\deg_{G^{i}}(u)\leq\textbf{{vol}}_{G^{i}}(D^{i}_{j})\leq\sum_{j\in[k_{i}]}\textbf{{vol}}_{G^{i}}(D^{i}_{j})\leq\tau\lambda/\phi,

and similarly for degGk(v)\deg_{G^{k}}(v). Therefore, by Lemma 3.8,

|w(GS)Ww(HS)|((L+1)τϕ)2ϵλ,\big{|}w(\partial_{G}S)-W\cdot w(\partial_{H}S)\big{|}\leq\left(\frac{(L+1)\tau}{\phi}\right)^{2}\cdot\epsilon\lambda,

as desired. ∎

The rest of Section 3.2 is dedicated to proving Lemma 3.8.

Expand out LG=eELeL_{G}=\sum_{e\in E}L_{e}, where LeL_{e} is the Laplacian of the graph consisting of the single edge ee of the same weight, so that 𝟙u¯TLe𝟙v¯{w(e),w(e)}\mathbbm{1}_{\overline{u}}^{T}L_{e}\mathbbm{1}_{\overline{v}}\in\{-w(e),w(e)\} if exactly one endpoint of ee is in u¯\overline{u} and exactly one endpoint of ee is in v¯\overline{v}, and 𝟙u¯TLe𝟙v¯=0\mathbbm{1}_{\overline{u}}^{T}L_{e}\mathbbm{1}_{\overline{v}}=0 otherwise. Let Eu¯,v¯,+E_{\overline{u},\overline{v},+} denote the edges eEe\in E with 𝟙u¯TLe𝟙v¯=w(e)\mathbbm{1}_{\overline{u}}^{T}L_{e}\mathbbm{1}_{\overline{v}}=w(e), and Eu¯,v¯,E_{\overline{u},\overline{v},-} denote those with 𝟙u¯TLe𝟙v¯=w(e)\mathbbm{1}_{\overline{u}}^{T}L_{e}\mathbbm{1}_{\overline{v}}=-w(e).

3.2.1 Random Sampling Procedure

Consider the Benzcur-Karger random sampling procedure, which we will de-randomize in this section. Let H^\widehat{H} be a subgraph of GG with each edge eEe\in E sampled independently with probability w(e)/Ww(e)/W, which is at most 11 by the assumption of Theorem 3.1. Intuitively, the parameter Wλ/f(n)W\geq\lambda/f(n) is selected so that with probability close to 11, (10) holds over all i,k,u,vi,k,u,v.

We now introduce our concentration bounds for the random sampling procedure, namely the classical multiplicative Chernoff bound. We state a form that includes bounds on the moment-generating function 𝔼[etX]\mathop{\mathbb{E}}[e^{tX}] obtained in the standard proof.

Lemma 3.10 (Multiplicative Chernoff bound).

Let X1,,XNX_{1},\ldots,X_{N} be independent random variables that take values in [0,1][0,1], and let X=i=1NXiX=\sum_{i=1}^{N}X_{i} and μ=𝔼[X]=i=1Npi\mu=\mathop{\mathbb{E}}[X]=\sum_{i=1}^{N}p_{i}. Fix a parameter δ\delta, and define

tu=ln(1+δ)andtl=ln(11δ).\displaystyle t^{u}=\ln(1+\delta)\qquad\text{and}\qquad t^{l}=\ln\left(\frac{1}{1-\delta}\right). (11)

Then, we have the following upper and lower tail bounds:

Pr[X>(1+δ)μ]\displaystyle\Pr[X>(1+\delta)\mu] etu(1+δ)μ𝔼[etuX]eδ2μ/3,\displaystyle\leq e^{-t^{u}(1+\delta)\mu}\mathop{\mathbb{E}}[e^{t^{u}X}]\leq e^{-\delta^{2}\mu/3}, (12)
Pr[X<(1δ)μ]\displaystyle\Pr[X<(1-\delta)\mu] etl(1δ)μ𝔼[etlX]eδ2μ/3.\displaystyle\leq e^{t^{l}(1-\delta)\mu}\mathop{\mathbb{E}}[e^{-t^{l}X}]\leq e^{-\delta^{2}\mu/3}. (13)

We now describe our de-randomization by pessimistic estimators. Let FEF\subseteq E be the set of edges for which a value Xe{0,1}X_{e}\in\{0,1\} has already been set, so that FF is initially \emptyset. For each i,ki,k, vertices uUi,vUku\in U^{i},v\in U^{k}, and sign {+,}\circ\in\{+,-\} such that Eu¯,v¯,E_{\overline{u},\overline{v},\circ}\neq\emptyset, we first define a “local” pessimistic estimator Φu¯,v¯,()\Phi_{\overline{u},\overline{v},\circ}(\cdot), which is a function on the set of pairs (e,Xe)(e,X_{e}) over all eFe\in F. The algorithm computes a 33-approximation λ~[λ,3λ]\widetilde{\lambda}\in[\lambda,3\lambda] to the mincut with the O~(m)\widetilde{O}(m)-time (2+ϵ)(2+\epsilon)-approximation algorithm of Matula [Mat93], and sets

μu¯,v¯,=w(Eu¯,v¯,)Wandδu¯,v¯,=ϵλ~6w(Eu¯,v¯,).\displaystyle\mu_{\overline{u},\overline{v},\circ}=\frac{w(E_{\overline{u},\overline{v},\circ})}{W}\qquad\text{and}\qquad\delta_{\overline{u},\overline{v},\circ}=\frac{\epsilon\widetilde{\lambda}}{6w(E_{\overline{u},\overline{v},\circ})}. (14)

Following (11), we define

tu¯,v¯,u=ln(1+δu¯,v¯,)andtu¯,v¯,l=ln(11δu¯,v¯,),\displaystyle t^{u}_{\overline{u},\overline{v},\circ}=\ln(1+\delta_{\overline{u},\overline{v},\circ})\qquad\text{and}\qquad t^{l}_{\overline{u},\overline{v},\circ}=\ln\left(\frac{1}{1-\delta_{\overline{u},\overline{v},\circ}}\right), (15)

and following the middle expressions (the moment-generating functions) in (12) and (13), we define

Φu¯,v¯,({(e,Xe):eF})=etu¯,v¯,u(1+δu¯,v¯,)μu¯,v¯,\displaystyle\Phi_{\overline{u},\overline{v},\circ}(\{(e,X_{e}):e\in F\})=e^{-t^{u}_{\overline{u},\overline{v},\circ}(1+\delta_{\overline{u},\overline{v},\circ})\mu_{\overline{u},\overline{v},\circ}} eEu¯,v¯,Fetu¯,v¯,uXeeEu¯,v¯,F𝔼[etu¯,v¯,uXe]\displaystyle\prod_{e\in E_{\overline{u},\overline{v},\circ}\cap F}e^{t^{u}_{\overline{u},\overline{v},\circ}X_{e}}\prod_{e\in E_{\overline{u},\overline{v},\circ}\setminus F}\mathop{\mathbb{E}}[e^{t^{u}_{\overline{u},\overline{v},\circ}X_{e}}]
+etu¯,v¯,l(1δu¯,v¯,)μu¯,v¯,\displaystyle+\;e^{t^{l}_{\overline{u},\overline{v},\circ}(1-\delta_{\overline{u},\overline{v},\circ})\mu_{\overline{u},\overline{v},\circ}} eEu¯,v¯,Fetu¯,v¯,lXeeEu¯,v¯,F𝔼[etu¯,v¯,lXe].\displaystyle\prod_{e\in E_{\overline{u},\overline{v},\circ}\cap F}e^{-t^{l}_{\overline{u},\overline{v},\circ}X_{e}}\prod_{e\in E_{\overline{u},\overline{v},\circ}\setminus F}\mathop{\mathbb{E}}[e^{-t^{l}_{\overline{u},\overline{v},\circ}X_{e}}].

Observe that if we are setting the value of XeX_{e^{\prime}} for a new edge eEu¯,v¯,Fe^{\prime}\in E_{\overline{u},\overline{v},\circ}\setminus F, then by linearity of expectation, there is an assignment Xe{0,1}X_{e^{\prime}}\in\{0,1\} for which Φu¯,v¯,()\Phi_{\overline{u},\overline{v},\circ}(\cdot) does not decrease:

Φu¯,v¯,({(e,Xe):eF}(e,Xe))Φu¯,v¯,({(e,Xe):eF}).\Phi_{\overline{u},\overline{v},\circ}(\{(e,X_{e}):e\in F\}\cup(e^{\prime},X_{e^{\prime}}))\leq\Phi_{\overline{u},\overline{v},\circ}(\{(e,X_{e}):e\in F\}).

Since the XeX_{e} terms are independent, we have that for any tt\in\mathbb{R} and EEE^{\prime}\subseteq E,

𝔼[eteEXe]=eE𝔼[etXe].\mathop{\mathbb{E}}\left[e^{t\sum_{e\in E^{\prime}}X_{e}}\right]=\displaystyle\prod\limits_{e\in E^{\prime}}\mathop{\mathbb{E}}[e^{tX_{e}}].

By the independence above and the second inequalities in (12) and (13), the initial “local” pessimistic estimator Φu¯,v¯,()\Phi_{\overline{u},\overline{v},\circ}(\emptyset) satisfies

Φu¯,v¯,()2exp(δu¯,v¯,2μu¯,v¯,3)\displaystyle\Phi_{\overline{u},\overline{v},\circ}(\emptyset)\leq 2\exp\left(-\frac{\delta_{\overline{u},\overline{v},\circ}^{2}\mu_{\overline{u},\overline{v},\circ}}{3}\right) =2exp((ϵλ~/(6w(Eu¯,v¯,)))2w(Eu¯,v¯,)/W3)\displaystyle=2\exp\left(-\frac{(\epsilon\widetilde{\lambda}/(6w(E_{\overline{u},\overline{v},\circ})))^{2}\cdot w(E_{\overline{u},\overline{v},\circ})/W\cdot}{3}\right)
=2exp(ϵλ~2108w(Eu¯,v¯,)W).\displaystyle=2\exp\left(-\frac{\epsilon\widetilde{\lambda}^{2}}{108w(E_{\overline{u},\overline{v},\circ})W}\right).

We would like the above expression to be less than 11. To upper bound w(Eu¯,v¯,)w(E_{\overline{u},\overline{v},\circ}), note first that every edge eEu¯,v¯,e\in E_{\overline{u},\overline{v},\circ} must, under the contraction from GG all the way to GiG^{i}, map to an edge incident to uu in GiG^{i}, which gives w(Eu¯,v¯,)degGi(u)w(E_{\overline{u},\overline{v},\circ})\leq\deg_{G^{i}}(u). Moreover, since degGi(u)τλ/ϕ\deg_{G^{i}}(u)\leq\tau\lambda/\phi by assumption, we have

w(Eu¯,v¯,)degGi(u)τλ/ϕ\displaystyle w(E_{\overline{u},\overline{v},\circ})\leq\deg_{G^{i}}(u)\leq\tau\lambda/\phi (16)

so that

Φu¯,v¯,()2exp(ϵλ~2108(τλ/ϕ)W)2exp(ϵλ2108(τλ/ϕ)W)=2exp(ϵϕλ108τW).\Phi_{\overline{u},\overline{v},\circ}(\emptyset)\leq 2\exp\left(-\frac{\epsilon\widetilde{\lambda}^{2}}{108(\tau\lambda/\phi)W}\right)\leq 2\exp\left(-\frac{\epsilon\lambda^{2}}{108(\tau\lambda/\phi)W}\right)=2\exp\left(-\frac{\epsilon\phi\lambda}{108\tau W}\right).

Assume that

Wϵϕλ108τln(16(L+1)2m),\displaystyle W\leq\frac{\epsilon\phi\lambda}{108\tau\ln\left(16(L+1)^{2}m\right)}, (17)

which satisfies the bounds in Lemma 3.8, so that

Φu¯,v¯,()2exp(ϵϕλ108τW)18(L+1)2m.\Phi_{\overline{u},\overline{v},\circ}(\emptyset)\leq 2\exp\left(-\frac{\epsilon\phi\lambda}{108\tau W}\right)\leq\frac{1}{8(L+1)^{2}m}.

Our actual, “global” pessimistic estimator Φ()\Phi(\cdot) is simply the sum of the “local” pessimistic estimators:

Φ({(e,Xe):eF})=i,k,uUi,vUk,{+,}Φu¯,v¯,({(e,Xe):eF}).\Phi(\{(e,X_{e}):e\in F\})=\sum_{\begin{subarray}{c}i,k,\\ u\in U^{i},v\in U^{k},\\ \circ\in\{+,-\}\end{subarray}}\Phi_{\overline{u},\overline{v},\circ}(\{(e,X_{e}):e\in F\}).

The initial pessimistic estimator Φ()\Phi(\emptyset) satisfies

Φ()=i,k,uUi,vUk,{+,}Φu¯,v¯,()i,k,uUi,vUk,{+,}18(L+1)2mClm.3.124(L+1)2m18(L+1)2m=12.\Phi(\emptyset)=\sum_{\begin{subarray}{c}i,k,\\ u\in U^{i},v\in U^{k},\\ \circ\in\{+,-\}\end{subarray}}\Phi_{\overline{u},\overline{v},\circ}(\emptyset)\leq\sum_{\begin{subarray}{c}i,k,\\ u\in U^{i},v\in U^{k},\\ \circ\in\{+,-\}\end{subarray}}\frac{1}{8(L+1)^{2}m}\stackrel{{\scriptstyle\text{Clm.}\ref{clm:for-each-edge}}}{{\leq}}4(L+1)^{2}m\cdot\frac{1}{8(L+1)^{2}m}=\frac{1}{2}.

Again, if we are setting the value of XfX_{f} for a new edge fEFf\in E\setminus F, then by linearity of expectation, there is an assignment Xf{0,1}X_{f}\in\{0,1\} for which Φ()\Phi(\cdot) does not decrease:

Φ({(e,Xe):eF}(f,Xf))Φ({(e,Xe):eF}).\Phi(\{(e,X_{e}):e\in F\}\cup(f,X_{f}))\leq\Phi(\{(e,X_{e}):e\in F\}).

Therefore, if we always select such an assignment XeX_{e}, then once we have iterated over all eEe\in E, we have

Φ({(e,Xe):eE})Φ()121.\displaystyle\Phi(\{(e,X_{e}):e\in E\})\leq\Phi(\emptyset)\leq\frac{1}{2}\leq 1. (18)

This means that for each i,k,uUi,vUki,k,u\in U^{i},v\in U^{k}, and sign {+,}\circ\in\{+,-\},

Φu¯,v¯,({(e,Xe):eE})=etu¯,v¯,u(1+δu¯,v¯,)μu¯,v¯,eEu¯,v¯,etu¯,v¯,uXe+etu¯,v¯,l(1δu¯,v¯,)μu¯,v¯,eEu¯,v¯,etu¯,v¯,lXe1.\Phi_{\overline{u},\overline{v},\circ}(\{(e,X_{e}):e\in E\})=e^{-t^{u}_{\overline{u},\overline{v},\circ}(1+\delta_{\overline{u},\overline{v},\circ})\mu_{\overline{u},\overline{v},\circ}}\prod_{e\in E_{\overline{u},\overline{v},\circ}}e^{t^{u}_{\overline{u},\overline{v},\circ}X_{e}}+\;e^{t^{l}_{\overline{u},\overline{v},\circ}(1-\delta_{\overline{u},\overline{v},\circ})\mu_{\overline{u},\overline{v},\circ}}\prod_{e\in E_{\overline{u},\overline{v},\circ}}e^{-t^{l}_{\overline{u},\overline{v},\circ}X_{e}}\leq 1.

In particular, each of the two terms is at most 11. Recalling from definition (14) that μu¯,v¯,=w(Eu¯,v¯,)/W\mu_{\overline{u},\overline{v},\circ}=w(E_{\overline{u},\overline{v},\circ})/W and δu¯,v¯,=ϵλ~/(6w(Eu¯,v¯,))\delta_{\overline{u},\overline{v},\circ}=\epsilon\widetilde{\lambda}/(6w(E_{\overline{u},\overline{v},\circ})), we have

eEu¯,v¯,Xe(1+δu¯,v¯,)μu¯,v¯,=w(Eu¯,v¯,)W+ϵλ~6W\sum_{e\in E_{\overline{u},\overline{v},\circ}}X_{e}\leq(1+\delta_{\overline{u},\overline{v},\circ})\mu_{\overline{u},\overline{v},\circ}=\frac{w(E_{\overline{u},\overline{v},\circ})}{W}+\frac{\epsilon\widetilde{\lambda}}{6W}

and

eEu¯,v¯,Xe(1δu¯,v¯,)μu¯,v¯,=w(Eu¯,v¯,)Wϵλ~6W.\sum_{e\in E_{\overline{u},\overline{v},\circ}}X_{e}\geq(1-\delta_{\overline{u},\overline{v},\circ})\mu_{\overline{u},\overline{v},\circ}=\frac{w(E_{\overline{u},\overline{v},\circ})}{W}-\frac{\epsilon\widetilde{\lambda}}{6W}.

Therefore,

|𝟙u¯TLG𝟙v¯W𝟙u¯TLH^𝟙v¯|{+,}|w(Eu¯,v¯,)WeEu¯,v¯,Xe|ϵλ~6+ϵλ~6=ϵλ~3ϵλ,\left|\mathbbm{1}^{T}_{\overline{u}}L_{G}\mathbbm{1}_{\overline{v}}-W\cdot\mathbbm{1}^{T}_{\overline{u}}L_{\widehat{H}}\mathbbm{1}_{\overline{v}}\right|\leq\sum_{\circ\in\{+,-\}}\left|w(E_{\overline{u},\overline{v},\circ})-W\cdot\sum_{e\in E_{\overline{u},\overline{v},\circ}}X_{e}\right|\leq\frac{\epsilon\widetilde{\lambda}}{6}+\frac{\epsilon\widetilde{\lambda}}{6}=\frac{\epsilon\widetilde{\lambda}}{3}\leq\epsilon\lambda,

fulfilling (10).

It remains to consider the running time. We first bound the number of i,k,u,vi,k,u,v such that either Eu¯,v¯,+E_{\overline{u},\overline{v},+}\neq\emptyset or Eu¯,v¯,E_{\overline{u},\overline{v},-}\neq\emptyset; the others are irrelevant since 𝟙u¯TLG𝟙v¯=𝟙u¯TLH^𝟙v¯=0\mathbbm{1}_{\overline{u}}^{T}L_{G}\mathbbm{1}_{\overline{v}}=\mathbbm{1}_{\overline{u}}^{T}L_{\widehat{H}}\mathbbm{1}_{\overline{v}}=0.

Claim 3.11.

For each pair of vertices x,yx,y, there are at most (L+1)2(L+1)^{2} many selections of i,ki,k and uUi,vUku\in U^{i},v\in U^{k} such that xu¯x\in\overline{u} and yv¯y\in\overline{v}.

Proof.

For each level ii, there is exactly one vertex uUiu\in U^{i} with xu¯x\in\overline{u}, and for each level kk, there is exactly one vertex vUkv\in U^{k} with yv¯y\in\overline{v}. This makes (L+1)2(L+1)^{2} many choices of i,ki,k total, and unique choices for u,vu,v given i,ki,k. ∎

Claim 3.12.

For each edge eEe\in E, there are at most 4(L+1)24(L+1)^{2} many selections of i,ki,k and uUi,vUku\in U^{i},v\in U^{k} such that eEu¯,v¯,+Eu¯,v¯,e\in E_{\overline{u},\overline{v},+}\cup E_{\overline{u},\overline{v},-}.

Proof.

If eEu¯,v¯,+Eu¯,v¯,e\in E_{\overline{u},\overline{v},+}\cup E_{\overline{u},\overline{v},-}, then exactly one endpoint of ee is in u¯\overline{u} and exactly one endpoint of ee is in v¯\overline{v}. There are four possibilities as to which endpoint is in u¯\overline{u} and which is in v¯\overline{v}, and for each, Claim 3.11 gives at most (L+1)2(L+1)^{2} choices. ∎

Claim 3.13.

There are at most 4(L+1)2m4(L+1)^{2}m many choices of i,k,u,vi,k,u,v such that either Eu¯,v¯,+E_{\overline{u},\overline{v},+}\neq\emptyset or Eu¯,v¯,E_{\overline{u},\overline{v},-}\neq\emptyset.

Proof.

For each such choice, charge it to an arbitrary edge (x,y)Eu¯,v¯,+Eu¯,v¯,(x,y)\in E_{\overline{u},\overline{v},+}\cup E_{\overline{u},\overline{v},-}. Each edge is charged at most 4(L+1)24(L+1)^{2} times by Claim 3.12, giving at most 4(L+1)2m4(L+1)^{2}m total charges. ∎

By Claim 3.12, each new edge eEFe\in E\setminus F is in at most 4(L+1)24(L+1)^{2} many sets Eu¯,v¯,E_{\overline{u},\overline{v},\circ}, and therefore affects at most 4(L+1)24(L+1)^{2} many terms Φu¯,v¯,({(e,Xe):eF})\Phi_{\overline{u},\overline{v},\circ}(\{(e,X_{e}):e\in F\}). The algorithm only needs to re-evaluate these terms with the new variable XeX_{e} set to 0 and with it set to 11, and take the one with the smaller new Φ()\Phi(\cdot). This takes O(L2)O(L^{2}) arithmetic operations.

How long do the arithmetic operations take? We compute each exponential in Φ()\Phi(\cdot) with clognc\log n bits of precision after the decimal point for some constant c>0c>0, which takes polylog(n)\textup{polylog}(n) time. Each one introduces an additive error of 1/nc1/n^{c}, and there are poly(n)\textup{poly}(n) exponential computations overall, for a total of 1/ncpoly(n)1/21/n^{c}\cdot\textup{poly}(n)\leq 1/2 error for a large enough c>0c>0. Factoring in this error, the inequality (18) instead becomes

Φ({(e,Xe):eE})Φ()+1212+12=1,\Phi(\{(e,X_{e}):e\in E\})\leq\Phi(\emptyset)+\frac{1}{2}\leq\frac{1}{2}+\frac{1}{2}=1,

so the rest of the bounds still hold.

This concludes the proof of Lemma 3.8.

3.2.2 Balanced Case

Similar to the expander case, we treat balanced cuts by “overlaying” a “lossy”, no(1)n^{o(1)}-approximate sparsifier of GG top of the graph H^\widehat{H} obtained from Lemma 3.9. In the expander case, this sparsifier was just another expander, but for general graphs, we need to do more work. At a high level, we compute an expander decomposition sequence, and on each level, we replace each of the expanders with a fixed expander (like in the expander case). Due to the technical proof and lack of novel ideas, we defer the proof to Appendix B.

Theorem 3.14.

Let GG be an weighted multigraph with mincut λ\lambda whose edges have weight at most O(λ)O(\lambda). For any parameters λ~[λ,3λ]\widetilde{\lambda}\in[\lambda,3\lambda] and Δ2O(logn)5/6\Delta\geq 2^{O(\log n)^{5/6}}, we can compute, in deterministic 2O(logn)5/6(loglogn)O(1)m+O(Δm)2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m+O(\Delta m) time, an unweighted multigraph HH such that WHW\cdot H is a γ\gamma-approximate cut sparsifier of GG, where γ2O(logn)5/6(loglogn)O(1)\gamma\leq 2^{O(\log n)^{5/6}(\log\log n)^{O(1)}} and W=λ~/ΔW=\widetilde{\lambda}/\Delta. (The graph HH does not need to be a subgraph of GG.) Moreover, the algorithm does not need to know the mincut value λ\lambda.

3.2.3 Combining Them Together

We now combine the unbalanced and balanced cases to prove Theorem 3.1, restated below.

See 3.1

Our high-level procedure is similar to the one from the expander case. For the τ\tau-unbalanced cuts, we use Lemma 3.9. For the balanced cuts, we show that their size must be much larger than λ\lambda, so that even on a γ\gamma-approximate weighted sparsifier guaranteed by Theorem 3.14, their weight is still much larger than λ\lambda. We then “overlay” the γ\gamma-approximate weighted sparsifier with a “light” enough weight onto the sparsifier of τ\tau-unbalanced cuts. The weight is light enough to barely affect the mincuts, but still large enough to force any balanced cut to increase by at least λ\lambda in weight.

Claim 3.15.

If a cut SS is balanced, then w(GS)βO(L)τλw(\partial_{G}S)\geq\beta^{O(L)}\tau\lambda.

Proof.

Consider the level ii for which j[ki]volGi(Dji)>τλ/ϕ\sum_{j\in[k_{i}]}\textbf{{vol}}_{G^{i}}(D^{i}_{j})>\tau\lambda/\phi. For each j[ki]j\in[k_{i}], we have

volGi(Dji)=volGi[Uji](Dji)+w(EGi(Dji,UiUji))\displaystyle\textbf{{vol}}_{G^{i}}(D^{i}_{j})=\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})) (7)1ϕw(Gi[Uji]Dji)+w(EGi(Dji,UiUji))\displaystyle\stackrel{{\scriptstyle(\ref{eq:Exp})}}{{\leq}}\frac{1}{\phi}w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
1ϕ(w(Gi[Uji]Dji)+w(EGi(Dji,UiUji)))\displaystyle\leq\frac{1}{\phi}\left(w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\right)
=1ϕw(GiDji),\displaystyle=\frac{1}{\phi}w(\partial_{G^{i}}D^{i}_{j}),

so summing over all j[ki]j\in[k_{i}],

j[ki]1ϕw(GiDji)j[ki]volGi(Dji)>τλϕ.\sum_{j\in[k_{i}]}\frac{1}{\phi}w(\partial_{G^{i}}D^{i}_{j})\geq\sum_{j\in[k_{i}]}\textbf{{vol}}_{G^{i}}(D^{i}_{j})>\frac{\tau\lambda}{\phi}.

By Lemma 3.4, it follows that

w(GS)βO(L)j[ki]w(GiDji)βO(L)τλ.w(\partial_{G}S)\geq\beta^{O(L)}\sum_{j\in[k_{i}]}w(\partial_{G}^{i}D^{i}_{j})\geq\beta^{O(L)}\tau\lambda.

Par. Value
λ\lambda Mincut of GG
λ~\widetilde{\lambda} 33-approximation of λ\lambda
ϵ\epsilon Given as input
rr (logn)1/6(\log n)^{1/6}
β\beta (logn)O(r4)(\log n)^{-O(r^{4})} from Theorem 3.2
ϕ\phi (logn)r5(\log n)^{-r^{5}}
LL O(lognr5)O(\frac{\log n}{r^{5}})
γ\gamma 2O(logn)5/6(loglogn)O(1)2^{O(\log n)^{5/6}(\log\log n)^{O(1)}} from Theorem 3.14
Δ\Delta 2Θ(logn)5/62^{\Theta(\log n)^{5/6}} from Theorem 3.14
τ\tau βcLγ2/ϵ\beta^{-cL}\gamma^{2}/\epsilon for large enough constant c>0c>0
ϵ\epsilon^{\prime} 12(ϕ(L+1)τ)2ϵ\frac{1}{2}(\frac{\phi}{(L+1)\tau})^{2}\epsilon
W^\widehat{W} min{Cϵϕλ~τln(Lm),λ~Δ}\min\{\frac{C\epsilon^{\prime}\phi\widetilde{\lambda}}{\tau\ln(Lm)},\frac{\widetilde{\lambda}}{\Delta}\} where C>0C>0 is the constant from Lemma 3.9
W~\widetilde{W} ϵ2γλ~Δ\frac{\epsilon}{2\gamma}\cdot\frac{\widetilde{\lambda}}{\Delta}
Figure 1: The parameters in the proof of Theorem 3.1.

We now set some of our parameters; see Figure 1 for a complete table of the parameters in our proof. For r:=(logn)1/6r:=(\log n)^{1/6}, let β:=(logn)O(r4)\beta:=(\log n)^{-O(r^{4})} and ϕ:=(logn)r5\phi:=(\log n)^{-r^{5}}, so that by Theorem 3.2, the total weight of inter-cluster edges, and therefore the total weight of the next graph in the expander decomposition sequence, shrinks by factor (logn)O(r4)ϕ=(logn)Ω(r5)(\log n)^{O(r^{4})}\phi=(\log n)^{-\Omega(r^{5})}. Since edge weights are assumed to be polynomially bounded, this shrinking can only happen O(lognr5)O(\frac{\log n}{r^{5}}) times, so LO(lognr5)L\leq O(\frac{\log n}{r^{5}}).

Let λ~[λ,3λ]\widetilde{\lambda}\in[\lambda,3\lambda] be a 33-approximation to the mincut, computable in O~(m)\widetilde{O}(m) time [Mat93], Let ϵ:=12(ϕ(L+1)τ)2ϵ\epsilon^{\prime}:=\frac{1}{2}(\frac{\phi}{(L+1)\tau})^{2}\epsilon for parameter τ\tau that we set later, and let H^\widehat{H} be the sparsifier of τ\tau-unbalanced cuts from Lemma 3.9 for this value of ϵ\epsilon^{\prime} (instead of ϵ\epsilon) and the following value of W^Cϵϕλτln(Lm)\widehat{W}\leq\frac{C\epsilon^{\prime}\phi\lambda}{\tau\ln(Lm)} (taking the place of WW):

W^:=min{Cϵϕλ~3τln(Lm),λ~Δ}=min{Ω(ϵϕ3λ~τ3L2ln(Lm)),λ~Δ}.\widehat{W}:=\min\left\{\frac{C\epsilon^{\prime}\phi\widetilde{\lambda}}{3\tau\ln(Lm)},\frac{\widetilde{\lambda}}{\Delta}\right\}=\min\left\{\Omega\left(\frac{\epsilon\phi^{3}\widetilde{\lambda}}{\tau^{3}L^{2}\ln(Lm)}\right),\frac{\widetilde{\lambda}}{\Delta}\right\}.

Let H~\widetilde{H} be the unweighted graph from Theorem 3.14 applied to λ~\widetilde{\lambda} and Δ\Delta, so that λ~/ΔH~\widetilde{\lambda}/\Delta\cdot\widetilde{H} is a γ\gamma-approximate cut sparsifier for γ:=2O(logn)5/6(loglogn)O(1)\gamma:=2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}. Define W~:=ϵ2γλ~Δ\widetilde{W}:=\frac{\epsilon}{2\gamma}\cdot\frac{\widetilde{\lambda}}{\Delta}, and let HH^{\prime} be the “union” of the graph H^\widehat{H} weighted by W^\widehat{W} and the graph H~\widetilde{H} weighted by W~\widetilde{W}. More formally, consider a weighted graph HH^{\prime} where each edge (u,v)(u,v) is weighted by W^wH^(u,v)+W~wH~(u,v)\widehat{W}\cdot w_{\widehat{H}}(u,v)+\widetilde{W}\cdot w_{\widetilde{H}}(u,v).

For an τ\tau-unbalanced cut S\partial S, the addition of the graph H~\widetilde{H} weighted by W~\widetilde{W} increases its weight by

W~w(H~S)=ϵ2γ(λΔw(H~S))ϵ2γγw(GS)=ϵ2w(GS),\widetilde{W}\cdot w(\partial_{\widetilde{H}}S)=\frac{\epsilon}{2\gamma}\cdot\left(\frac{\lambda}{\Delta}w(\partial_{\widetilde{H}}S)\right)\leq\frac{\epsilon}{2\gamma}\cdot\gamma w(\partial_{G}S)=\frac{\epsilon}{2}w(\partial_{G}S),

so that

|w(GS)(W^w(H^S)+W~w(H~S))|\displaystyle\left|w(\partial_{G}S)-\left(\widehat{W}\cdot w(\partial_{\widehat{H}}S)+\widetilde{W}\cdot w(\partial_{\widetilde{H}}S)\right)\right| |w(GS)W^w(H^S)|+W~w(H~S)\displaystyle\leq\big{|}w(\partial_{G}S)-\widehat{W}\cdot w(\partial_{\widehat{H}}S)\big{|}+\widetilde{W}\cdot w(\partial_{\widetilde{H}}S^{*})
((L+1)τϕ)2ϵλ+ϵ2w(GS)\displaystyle\leq\left(\frac{(L+1)\tau}{\phi}\right)^{2}\cdot\epsilon^{\prime}\lambda+\frac{\epsilon}{2}w(\partial_{G}S)
=ϵλ2+ϵ2w(GS)\displaystyle=\frac{\epsilon\lambda}{2}+\frac{\epsilon}{2}w(\partial_{G}S)
ϵw(GS).\displaystyle\leq\epsilon w(\partial_{G}S).

In particular, any τ\tau-unbalanced cut satisfies

(1ϵ)λW^w(H^S)+W~w(H~S)(1+ϵ)λ.\displaystyle(1-\epsilon)\lambda\leq\widehat{W}\cdot w(\partial_{\widehat{H}}S)+\widetilde{W}\cdot w(\partial_{\widetilde{H}}S)\leq(1+\epsilon)\lambda. (19)

Next, we show that all balanced cuts have weight at least λ\lambda in the graph H~\widetilde{H} weighted by W~\widetilde{W}. This is where we finally set τ:=βcLγ2/ϵ\tau:=\beta^{-cL}\gamma^{2}/\epsilon for large enough constant c>0c>0. For a balanced cut SS,

W~w(H~S)=ϵ2γ(λΔw(H~S))ϵ2γ(1γw(GS))Clm.3.15ϵγ2βO(L)τλλ.\widetilde{W}\cdot w(\partial_{\widetilde{H}}S)=\frac{\epsilon}{2\gamma}\cdot\left(\frac{\lambda}{\Delta}w(\partial_{\widetilde{H}}S)\right)\geq\frac{\epsilon}{2\gamma}\cdot\left(\frac{1}{\gamma}w(\partial_{G}S)\right)\stackrel{{\scriptstyle\text{Clm.}\ref{clm:bal}}}{{\geq}}\frac{\epsilon}{\gamma^{2}}\cdot\beta^{O(L)}\tau\lambda\geq\lambda.

Moreover, by Claim 3.6 for this value of τβO(L)\tau\geq\beta^{-O(L)}, the mincut S\partial S^{*} is τ\tau-unbalanced, and therefore has weight at least (1ϵ)λ(1-\epsilon)\lambda in HH^{\prime} by (19).

Therefore, HH^{\prime} preserves the mincut up to factor ϵ\epsilon and has mincut at least (1ϵ)λ(1-\epsilon)\lambda. It remains to make all edge weights the same on this sparsifier. Since W~=ϵ2γλ~Δ\widetilde{W}=\frac{\epsilon}{2\gamma}\cdot\frac{\widetilde{\lambda}}{\Delta} and the only requirement for Δ\Delta from Theorem 3.14 is that Δ2O(logn)5/6\Delta\geq 2^{O(\log n)^{5/6}}, we can increase or decrease Δ\Delta by a constant factor until either W~/W^\widetilde{W}/\widehat{W} or W^/W~\widehat{W}/\widetilde{W} is an integer. Then, we can let W:=min{W^,W~}W:=\min\{\widehat{W},\widetilde{W}\} and define the unweighted graph HH so that #H(u,v)=wH(u,v)/W\#_{H}(u,v)=w_{H^{\prime}}(u,v)/W for all u,vVu,v\in V. Therefore, our final weight WW is

W=min{W^,W~}\displaystyle W=\min\{\widehat{W},\widetilde{W}\} =min{Ω(ϵϕ3λ~τ3L2ln(Lm)),λ~Δ,ϵ2γλ~Δ}\displaystyle=\min\left\{\Omega\left(\frac{\epsilon\phi^{3}\widetilde{\lambda}}{\tau^{3}L^{2}\ln(Lm)}\right),\frac{\widetilde{\lambda}}{\Delta},\frac{\epsilon}{2\gamma}\cdot\frac{\widetilde{\lambda}}{\Delta}\right\}
ϵ42O(logn)5/6(loglogn)O(1)λ,\displaystyle\geq\epsilon^{4}2^{-O(\log n)^{5/6}(\log\log n)^{O(1)}}\lambda,

so we can set f(n):=2O(logn)5/6(loglogn)O(1)f(n):=2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}, as desired.

Finally, we bound the running time. The expander decomposition sequence (Theorem 3.2) takes time m1+O(1/r)+O~(m/ϕ2)m^{1+O(1/r)}+\widetilde{O}(m/\phi^{2}), the unbalanced case (Theorem 3.2) takes time O~(L2m)\widetilde{O}(L^{2}m), and the balanced case takes time 2O(logn)5/6(loglogn)O(1)m2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m. Altogether, the total is 2O(logn)5/6(loglogn)O(1)m2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m, which concludes the proof of Theorem 3.1.

3.3 Removing the Maximum Weight Assumption

Let f(n)=2O(logn)5/6(loglogn)O(1)f(n)=2^{O(\log n)^{5/6}(\log\log n)^{O(1)}} be the function from Theorem 3.1. In this section, we show how to use Theorem 3.1, which assumes that the maximum edge weight in GG is at most ϵ4λ/f(n)\epsilon^{4}\lambda/f(n), to prove Theorem 1.5, which makes no assumption on edge weights.

First, we show that we can assume without loss of generality that the maximum edge weight in GG is at most 3λ3\lambda. To see why, the algorithm can first compute a 33-approximation λ~[λ,3λ]\widetilde{\lambda}\in[\lambda,3\lambda] to the mincut with the O~(m)\widetilde{O}(m)-time (2+ϵ)(2+\epsilon)-approximation algorithm of Matula [Mat93], and for each edge in GG with weight more than λ~\widetilde{\lambda}, reduce its weight to λ~\widetilde{\lambda}. Let the resulting graph be G~\widetilde{G}. We now claim the following:

Claim 3.16.

Suppose an unweighted graph HH and some weight WW satisfy the two properties of Theorem 1.5 for G~\widetilde{G}. Then, they also satisfy the two properties of Theorem 1.5 for GG.

Proof.

The only cuts that change value between GG and G~\widetilde{G} are those with an edge of weight more than λ~\widetilde{\lambda}, which means their value must be greater than λ~λ\widetilde{\lambda}\geq\lambda. In particular, since GG and G~\widetilde{G} have the same mincuts and the same mincut values, both properties of Theorem 1.5 also hold when the input graph is GG. ∎

For the rest of the proof, we work with G~\widetilde{G} instead of GG. Define W~:=ϵ4λ~/(3f(n))\widetilde{W}:=\epsilon^{4}\widetilde{\lambda}/(3f(n)), which satisfies W~ϵ4λ/f(n)\widetilde{W}\leq\epsilon^{4}\lambda/f(n). For each edge ee in G~\widetilde{G}, split it into w(e)/W~\lceil w(e)/\widetilde{W}\rceil parallel edges of weight at most W~\widetilde{W} each, whose sum of weights equals w(e)w(e); let the resulting graph be G^\widehat{G}. Apply Theorem 3.1 on G^\widehat{G}, which returns an unweighted graph HH and weight Wϵ4λ/f(n)W\geq\epsilon^{4}\lambda/f(n) such that the two properties of Theorem 1.5 hold for G^\widehat{G}. Clearly, the cuts are the same in G~\widetilde{G} and G^\widehat{G}: we have w(G~S)=w(G^S)w(\partial_{\widetilde{G}}S)=w(\partial_{\widehat{G}}S) for all SVS\subseteq V. Therefore, the two properties also hold for G^\widehat{G}, as desired.

We now bound the size of GG^{\prime} and the running time. Since w(e)λ~w(e)\leq\widetilde{\lambda}, we have w(e)/W~3f(n)/ϵ4\lceil w(e)/\widetilde{W}\rceil\leq\lceil 3f(n)/\epsilon^{4}\rceil, so each edge splits into at most O(f(n)/ϵ4)O(f(n)/\epsilon^{4}) edges and the total number of edges is m^O(f(n)/ϵ4)m\widehat{m}\leq O(f(n)/\epsilon^{4})\cdot m. Therefore, Theorem 3.1 takes time 2O(logn)5/6(loglogn)O(1)m^=ϵ42O(logn)5/6(loglogn)O(1)m2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}\widehat{m}=\epsilon^{-4}2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m, concluding the proof of Theorem 1.5.

Acknowledgements

I am indebted to Sivakanth Gopi, Janardhan Kulkarni, Jakub Tarnawski, and Sam Wong for their supervision and encouragement on this project while I was a research intern at Microsoft Research, as well as providing valuable feedback on the manuscript. I also thank Thatchaphol Saranurak for introducing me to the boundary-linked expander decomposition framework [GRST20].

References

  • [CGL+19] Julia Chuzhoy, Yu Gao, Jason Li, Danupon Nanongkai, Richard Peng, and Thatchaphol Saranurak. A deterministic algorithm for balanced cut with applications to dynamic connectivity, flows, and beyond. arXiv preprint arXiv:1910.08025, 2019.
  • [Gab95] Harold N Gabow. A matroid approach to finding edge connectivity and packing arborescences. Journal of Computer and System Sciences, 50(2):259–273, 1995.
  • [GH61] Ralph E Gomory and Tien Chung Hu. Multi-terminal network flows. Journal of the Society for Industrial and Applied Mathematics, 9(4):551–570, 1961.
  • [GR98] Andrew V Goldberg and Satish Rao. Beyond the flow decomposition barrier. Journal of the ACM (JACM), 45(5):783–797, 1998.
  • [GRST20] Gramoz Goranci, Harald Räcke, Thatchaphol Saranurak, and Zihan Tan. The expander hierarchy and its applications to dynamic graph algorithms. arXiv preprint arXiv:2005.02369, 2020.
  • [GT88] Andrew V. Goldberg and Robert Endre Tarjan. A new approach to the maximum-flow problem. J. ACM, 35(4):921–940, 1988.
  • [HO92] Jianxiu Hao and James B Orlin. A faster algorithm for finding the minimum cut in a graph. In Proceedings of the third annual ACM-SIAM symposium on Discrete algorithms, pages 165–174. Society for Industrial and Applied Mathematics, 1992.
  • [HRW17] Monika Henzinger, Satish Rao, and Di Wang. Local flow partitioning for faster edge connectivity. In Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2017, Barcelona, Spain, Hotel Porta Fira, January 16-19, pages 1919–1938, 2017.
  • [Kar93] David R Karger. Global min-cuts in rnc, and other ramifications of a simple min-cut algorithm. In SODA, volume 93, pages 21–30, 1993.
  • [Kar00] David R. Karger. Minimum cuts in near-linear time. J. ACM, 47(1):46–76, 2000.
  • [KS96] David R Karger and Clifford Stein. A new approach to the minimum cut problem. Journal of the ACM (JACM), 43(4):601–640, 1996.
  • [KT18] Ken-ichi Kawarabayashi and Mikkel Thorup. Deterministic edge connectivity in near-linear time. Journal of the ACM (JACM), 66(1):1–50, 2018.
  • [LP20] Jason Li and Debmalya Panigrahi. Deterministic min-cut in poly-logarithmic max-flows. In FOCS, 2020.
  • [LS20] Yang P Liu and Aaron Sidford. Faster divergence maximization for faster maximum flow. arXiv preprint arXiv:2003.08929, 2020.
  • [LS21] Jason Li and Thatchaphol Saranurak. Deterministic weighted expander decomposition in almost-linear time, 2021.
  • [Mat93] David W Matula. A linear time 2+ ε\varepsilon approximation algorithm for edge connectivity. In Proceedings of the fourth annual ACM-SIAM Symposium on Discrete algorithms, pages 500–504, 1993.
  • [NI92a] Hiroshi Nagamochi and Toshihide Ibaraki. Computing edge-connectivity in multigraphs and capacitated graphs. SIAM Journal on Discrete Mathematics, 5(1):54–66, 1992.
  • [NI92b] Hiroshi Nagamochi and Toshihide Ibaraki. A linear-time algorithm for finding a sparse k-connected spanning subgraph of a k-connected graph. Algorithmica, 7(5&6):583–596, 1992.
  • [Sar21] Thatchaphol Saranurak. A simple deterministic algorithm for edge connectivity. In Symposium on Simplicity in Algorithms (SOSA), pages 80–85. SIAM, 2021.
  • [ST04] Daniel A. Spielman and Shang-Hua Teng. Nearly-linear time algorithms for graph partitioning, graph sparsification, and solving linear systems. In STOC, pages 81–90. ACM, 2004.
  • [SW97] Mechthild Stoer and Frank Wagner. A simple min-cut algorithm. Journal of the ACM (JACM), 44(4):585–591, 1997.
  • [SW19] Thatchaphol Saranurak and Di Wang. Expander decomposition and pruning: Faster, stronger, and simpler. In SODA, pages 2616–2635. SIAM, 2019.

Appendix A Boundary-Linked Expander Decomposition

In this section, we prove Theorem 3.2 assuming the subroutine WeightedBalCutPrune from [LS21]. Our proof is directly modeled off of the proof of Corollary 6.1 of [CGL+19] and the proof of Theorem 4.5 of [GRST20], so we claim no novelty in this section.

We will work with weighted multigraphs with self-loops, and we re-define the degree deg(v)\deg(v) to mean w(({v}))w(\partial(\{v\})) plus the total weight of all self-loops at vertex vv. All other definitions that depend on deg(v)\deg(v), such as vol(S)\textbf{{vol}}(S) and Φ(G)\Phi(G), are also affected.

Given a weighted graph G=(V,E)G=(V,E), a parameter r>0r>0, and a subset AVA\subseteq V, define G{A}rG\{A\}^{r} as the graph G[A]G[A] with the following self-loops attached: for each edge eE(A,VA)e\in E(A,V\setminus A) with endpoint vAv\in A, add a self-loop at vv of weight rw(e)r\cdot w(e). The following observation is immediate by definition:

Observation A.1.

For any graph G=(V,E)G=(V,E) and subset AVA\subseteq V, Property (1) of Theorem 3.2 holds for Vi=AV_{i}=A iff G[A]β/ϕG[A]^{\beta/\phi} is a ϕ\phi-expander.

We now define the WeightedBalCutPrune problem from [LS21]555 Their definition is more general and takes in a demand vector dV\textbf{{d}}\in\mathbb{R}^{V} on the vertices; we are simply restricting ourselves to d(v)=deg(v)\textbf{{d}}(v)=\deg(v) for all vVv\in V, which gives our definition. and their algorithm.

Definition A.2 (WeightedBalCutPrune problem, Definition 2.3 of [LS21]).

The input to the α\alpha-approximate WeightedBalCutPrune problem is a graph G=(V,E)G=(V,E), a conductance parameter 0<ϕ10<\phi\leq 1, and an approximation factor α\alpha. The goal is to compute a cut (A,B)(A,B) in GG, with wG(A,B)αϕvol(B)w_{G}(A,B)\leq\alpha\phi\cdot\textbf{{vol}}(B), such that one of the following holds: either

  1. 1.

    (Cut) vol(A),vol(B)vol(V)/3\textbf{{vol}}(A),\textbf{{vol}}(B)\geq\textbf{{vol}}(V)/3; or

  2. 2.

    (Prune) vol(A)vol(V)/2\textbf{{vol}}(A)\geq\textbf{{vol}}(V)/2, and Φ(G[A])ϕ\Phi(G[A])\geq\phi.

Theorem A.3 (WeightedBalCutPrune algorithm, Theorem 2.4 of [LS21]).

There is a deterministic algorithm that, given a graph G=(V,E)G=(V,E) with mm edges and polynomially bounded edge weights, and parameters 0<ψ10<\psi\leq 1 and r1r\geq 1, solves the (logn)O(r4)(\log n)^{O(r^{4})}-approximate WeightedBalCutPrune problem in time m1+O(1/r)m^{1+O(1/r)}.

The (Prune) case requires the additional trimming step described in the lemma below. While [GRST20] prove it for unweighted graphs only, the algorithm translates directly to the weighted case;666In particular, the core subroutine, called Unit-Flow in [SW19], is based on the push-relabel max-flow algorithm, which works on both unweighted and weighted graphs. see, for example, Theorem 4.2 of [SW19].

Lemma A.4 (Trimming, Lemmas 4.9 and 4.10 of [GRST20]).

Given a weighted graph G=(V,E)G=(V,E) and subset AVA\subseteq V such that G{A}G\{A\} is an 8ϕ8\phi-expander and w(EG(A,VA))ϕ16volG(A)w(E_{G}(A,V\setminus A))\leq\frac{\phi}{16}\textbf{{vol}}_{G}(A), we can compute a “pruned” set PAP\subseteq A in deterministic O~(m/ϕ2)\widetilde{O}(m/\phi^{2}) time with the following properties:

  1. 1.

    volG(P)4ϕw(EG(A,VA))\textbf{{vol}}_{G}(P)\leq\frac{4}{\phi}w(E_{G}(A,V\setminus A)),

  2. 2.

    w(EG(A,VA))2w(EG(A,VA))w(E_{G}(A^{\prime},V\setminus A^{\prime}))\leq 2w(E_{G}(A,V\setminus A)) where A:=APA^{\prime}:=A\setminus P, and

  3. 3.

    G{A}1/(8ϕ)G\{A^{\prime}\}^{1/(8\phi)} is a ϕ\phi-expander.

We now prove Theorem 3.2 assuming Theorem A.3. Our proof is copied almost ad verbatim from the proof of Corollary 6.1 of [CGL+19] on expander decompositions, with the necessary changes to prove the additional boundary-linked property.

We maintain a collection \mathcal{H} of vertex-disjoint graphs that we call clusters, which are subgraphs of GG with some additional self-loops. The set \mathcal{H} of clusters is partitioned into two subsets, set A\mathcal{H}^{A} of active clusters, and set I\mathcal{H}^{I} of inactive clusters. We ensure that each inactive cluster HIH\in\mathcal{H}^{I} is a ϕ\phi-expander. We also maintain a set EE^{\prime} of “deleted” edges, that are not contained in any cluster in \mathcal{H}. At the beginning of the algorithm, we let =A={G}\mathcal{H}=\mathcal{H}^{A}=\{G\}, I=\mathcal{H}^{I}=\emptyset, and E=E^{\prime}=\emptyset. The algorithm proceeds as long as A\mathcal{H}^{A}\neq\emptyset, and consists of iterations. Let α=(logn)O(r4)\alpha=(\log n)^{O(r^{4})} be the approximation factor from Theorem A.3.

In every iteration, we apply the algorithm from Theorem A.3 to every graph HAH\in\mathcal{H}^{A}, with the same parameters α\alpha, rr, and ϕ\phi. Let UU be the vertices of HH. Consider the cut (A,B)(A,B) in HH that the algorithm returns, with

w(EH(A,B))αϕvol(U)ϵvol(U)clogn.\displaystyle w(E_{H}(A,B))\leq\alpha\phi\cdot\textbf{{vol}}(U)\leq\frac{\epsilon\cdot\textbf{{vol}}(U)}{c\log n}. (20)

We add the edges of EH(A,B)E_{H}(A,B) to set EE^{\prime}.

If volH(B)vol(U)32α\textbf{{vol}}_{H}(B)\geq\frac{\textbf{{vol}}(U)}{32\alpha}, then we replace HH with H{A}1/(α2ϕlogn)H\{A\}^{1/(\alpha^{2}\phi\log n)} and H{B}1/(α2ϕlogn)H\{B\}^{1/(\alpha^{2}\phi\log n)} in \mathcal{H} and in A\mathcal{H}^{A}. Note that the self-loops add a total volume of

1α2ϕlognw(EH(A,B))1α2ϕlognαϕvol(U)=1αlognvol(U).\displaystyle\frac{1}{\alpha^{2}\phi\log n}\cdot w(E_{H}(A,B))\leq\frac{1}{\alpha^{2}\phi\log n}\cdot\alpha\phi\,\textbf{{vol}}(U)=\frac{1}{\alpha\log n}\textbf{{vol}}(U). (21)

Otherwise, if volH(B)<vol(U)32αvol(U)/3\textbf{{vol}}_{H}(B)<\frac{\textbf{{vol}}(U)}{32\alpha}\leq\textbf{{vol}}(U)/3, then we must be in the (Prune) case, which means that volH(A)vol(U)/2\textbf{{vol}}_{H}(A)\geq\textbf{{vol}}(U)/2 and graph H{A}1/(8ϕ)H\{A\}^{1/(8\phi)} has conductance at least ϕ\phi. Since

w(EH(A,B))αϕvolH(B)ϕ32vol(U)ϕ16vol(A),w(E_{H}(A,B))\leq\alpha\phi\cdot\textbf{{vol}}_{H}(B)\leq\frac{\phi}{32}\textbf{{vol}}(U)\leq\frac{\phi}{16}\textbf{{vol}}(A),

we can call Lemma A.4 on AA to obtain a pruned set PAP\subseteq A such that

volH(P)4ϕw(EH(A,B))18vol(U)\textbf{{vol}}_{H}(P)\leq\frac{4}{\phi}w(E_{H}(A,B))\leq\frac{1}{8}\textbf{{vol}}(U)

and

w(EH(A,UA))2w(EH(A,B))ϕ8vol(A)w(E_{H}(A^{\prime},U\setminus A^{\prime}))\leq 2w(E_{H}(A,B))\leq\frac{\phi}{8}\textbf{{vol}}(A)

for A:=APA^{\prime}:=A\setminus P, and H{A}1/(8ϕ)H\{A^{\prime}\}^{1/(8\phi)} is a ϕ\phi-expander. Add the edges of EH(A,UA)E_{H}(A^{\prime},U\setminus A^{\prime}) to EE^{\prime}, remove HH from \mathcal{H} and A\mathcal{H}^{A}, add H{A}1/(8ϕ)H\{A^{\prime}\}^{1/(8\phi)} to \mathcal{H} and I\mathcal{H}^{I}, and add H{BP}1/(8ϕ)H\{B\cup P\}^{1/(8\phi)} to \mathcal{H} and A\mathcal{H}^{A}. Observe that

volH(BP)=volH(B)+volH(P)12volH(U)+18volH(U)58vol(U).\textbf{{vol}}_{H}(B\cup P)=\textbf{{vol}}_{H}(B)+\textbf{{vol}}_{H}(P)\leq\frac{1}{2}\textbf{{vol}}_{H}(U)+\frac{1}{8}\textbf{{vol}}_{H}(U)\leq\frac{5}{8}\textbf{{vol}}(U).

When the algorithm terminates, A=\mathcal{H}^{A}=\emptyset, and so every graph in \mathcal{H} has conductance at least ϕ\phi. Notice that in every iteration, the maximum volume of a graph in A\mathcal{H}^{A} is at most a factor (1132α)(1-\frac{1}{32\alpha}) of what it was before. Since edge weights are polynomially bounded, the number of iterations is at most O(αlogn)O(\alpha\log n). On each iteration, the total volume of graphs in A\mathcal{H}^{A} increases by at most factor 1+2αlogn1+\frac{2}{\alpha\log n} factor due to the self-loops added in (21), so the total volume of all HH\in\mathcal{H} at the end is at most a constant factor of the initial volume volG(V)\textbf{{vol}}_{G}(V).

The output of the algorithm is the partition of VV induced by the vertex sets of HH\in\mathcal{H}, so the inter-cluster edges is a subset of EE^{\prime}. It is easy to verify by (20) that the total weight of edges added to set EE^{\prime} in every iteration is at most αϕ\alpha\phi times the total volume of graphs in A\mathcal{H}^{A} at the beginning of that iteration, which is O(volG(V))O(\textbf{{vol}}_{G}(V)). Over all O(αlogn)O(\alpha\log n) iterations, the total weight of EE^{\prime} is O(αlogn)αϕvolG(V)(logn)O(r4)ϕvolG(V)O(\alpha\log n)\cdot\alpha\phi\,\textbf{{vol}}_{G}(V)\leq(\log n)^{O(r^{4})}\phi\textbf{{vol}}_{G}(V), fulfilling property (2) of a boundary-linked expander decomposition.

It remains to show that for each graph HIH\in\mathcal{H}^{I}, its vertex set UU satisfies the boundary-linked ϕ\phi-expander property (1) of Theorem 3.2. For each boundary edge eEG(U,VU)e\in E_{G}(U,V\setminus U), it was created at some iteration where we either added 1α2ϕlogn\frac{1}{\alpha^{2}\phi\log n} self-loops or 18ϕ\frac{1}{8\phi} self-loops, so G[U]min{1/(α2ϕlogn),1/(8ϕ)}G[U]^{\min\{1/(\alpha^{2}\phi\log n),1/(8\phi)\}} is a subgraph of HH. Since HH is a ϕ\phi-expander, so is G[U]min{1/(α2ϕlogn),1/(8ϕ)}G[U]^{\min\{1/(\alpha^{2}\phi\log n),1/(8\phi)\}}, and property (1) for β:=min{1/α2,1/8}\beta:=\min\{1/\alpha^{2},1/8\} follows by Observation A.1.

It remains to analyze the running time of the algorithm. The running time of a single iteration is bounded by O(m1+O(1/r))+O~(m/ϕ2)O(m^{1+O(1/r)})+\widetilde{O}(m/\phi^{2}). Since the total number of iterations is bounded by O(logn)O(\log n), the total running time is the same, asymptotically.

Appendix B Lossy Unweighted Sparsifier

In this section, we prove Theorem 3.14, restated below.

See 3.14

We will work with weighted multigraphs with self-loops, and we re-define the degree deg(v)\deg(v) to mean w(({v}))w(\partial(\{v\})) plus the total weight of all self-loops at vertex vv. All other definitions that depend on deg(v)\deg(v), such as vol(S)\textbf{{vol}}(S) and Φ(G)\Phi(G), are also affected.

The construction of the sparsifier HH is recursive. The original input is graph G=G0G=G^{0}, and let the input graph on level i0i\geq 0 of the recursion be GiG^{i}, with UiU^{i} as its vertex set. Let U1i,U2i,U^{i}_{1},U^{i}_{2},\ldots be an expander decomposition of GiG^{i}, and let Gi+1G^{i+1} be the graph with each set UjiU^{i}_{j} contracted to a single vertex uji+1u^{i+1}_{j}. If Gi+1G^{i+1} has more than one vertex, recursively compute a sparsifier on Gi+1G^{i+1}, which still has mincut at least λ\lambda, and let the sparsifier be Hi+1H^{i+1}. For each edge (uji+1,uki+1)(u^{i+1}_{j},u^{i+1}_{k}) in Hi+1H^{i+1}, we select a vertex xUjix\in U^{i}_{j} and yUkiy\in U^{i}_{k} and add edge (x,y)(x,y) to an initially empty graph H0iH^{i}_{0} on UiU^{i}. We do this in a way that each vertex vUiv\in U^{i} is incident to at most degHi+1(uji+1)w(EGi(v,UiUji))degGi+1(uji+1)\left\lceil\deg_{H^{i+1}}(u^{i+1}_{j})\cdot\frac{w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j}))}{\deg_{G^{i+1}}(u^{i+1}_{j})}\right\rceil many edges. Since vUjiw(EGi(v,UiUji))=degGi+1(uji+1)\sum_{v\in U^{i}_{j}}w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j}))=\deg_{G^{i+1}}(u^{i+1}_{j}), this is always possible by an averaging argument. Next, for each cluster UjiU^{i}_{j}, we compute an Ω(1)\Omega(1)-expander multigraph HjiH^{i}_{j} (possibly with self-loops) on the vertices UjiU^{i}_{j} such that for all vUjiv\in U^{i}_{j},

degGi(v)WdegHji(v)9degGi(v).\displaystyle\deg_{G^{i}}(v)\leq W\cdot\deg_{H^{i}_{j}}(v)\leq 9\deg_{G^{i}}(v). (22)

This can be done by using the lemma below with d(v)=degGi(v)/Wλ/W1d(v)=\deg_{G^{i}}(v)/W\geq\lambda/W\geq 1. The running time is at most O(vUidegGi(v)/W)O(eEw(e)/W)O(\sum_{v\in U_{i}}\deg_{G^{i}}(v)/W)\leq O(\sum_{e\in E}w(e)/W), which is O(mλ/W)=O(Δm)O(m\lambda/W)=O(\Delta m) since by assumption, all edges in GG have weight at most O(λ)O(\lambda), and W=λ~/Δ=Θ(λ/Δ)W=\widetilde{\lambda}/\Delta=\Theta(\lambda/\Delta).

Lemma B.1.

Given a vertex set VV and real numbers d(v)1:vVd(v)\geq 1:v\in V, there exists a universal constant C0C_{0} such that we can construct, in O(vVd(v))O(\sum_{v\in V}d(v)) time, an Ω(1)\Omega(1)-expander multigraph HH on VV (possibly with self-loops) such that for all vVv\in V,

d(v)degH(v)9d(v).d(v)\leq\deg_{H}(v)\leq 9d(v).
Proof.

We use the following theorem of [CGL+19]:

Theorem B.2.

There is a constant α0>0\alpha_{0}>0 and a deterministic algorithm that, given an integer n>1n>1, in time O(n)O(n) constructs a graph HnH_{n} with |V(Hn)|=n|V(H_{n})|=n, such that HnH_{n} is an α0\alpha_{0}-expander, and every vertex in HnH_{n} has degree at most 99.

Let n=vVd(v)n=\sum_{v\in V}d(v), and let HnH_{n} be the constructed graph on vertex set VnV_{n}. Partition VnV_{n} arbitrarily into subsets Uv:vVU_{v}:v\in V such that |Uv|=d(v)|U_{v}|=d(v) for each vVv\in V. Let HH be the graph HnH_{n} with each set UvU_{v} contracted to a single vertex vv, keeping self-loops, so that degH(v)=volHn(Uv)\deg_{H}(v)=\textbf{{vol}}_{H_{n}}(U_{v}). It is not hard to see that expansion does not decrease upon contraction, so HH is still an Ω(1)\Omega(1)-expander. We can bound the degrees degH(v)\deg_{H}(v) as

d(v)=|Uv|volHn(Uv)=degH(v)=volHn(Uv)9|Uv|=9d(v).d(v)=|U_{v}|\leq\textbf{{vol}}_{H_{n}}(U_{v})=\deg_{H}(v)=\textbf{{vol}}_{H_{n}}(U_{v})\leq 9|U_{v}|=9d(v).

The final sparsifier HiH^{i} is H0iH1iH2iH^{i}_{0}\cup H^{i}_{1}\cup H^{i}_{2}\cup\cdots. This concludes the construction of sparsifier HiH^{i}. (We keep the self-loops, even though they serve no purpose for the sparsifier’s guarantees, because we find that including them simplifies the analysis.) Note that this recursive algorithm implicitly constructs an expander sequence G0,G1,G2,,GLG^{0},G^{1},G^{2},\ldots,G^{L} of GG over its recursive calls.

Fix a subset SUi\emptyset\subsetneq S\subsetneq U^{i}, let SiS^{i} be the canonical decomposition sequence of SS, and let DjiD^{i}_{j} be constructed as before, so that they satisfy (7) and (8) for all i,ji,j.

Claim B.3.

For all ii and all vUiv\in U^{i},

degGi(v)WdegHi(v)10(L+1)degGi(v).\deg_{G^{i}}(v)\leq W\cdot\deg_{H^{i}}(v)\leq 10(L+1)\cdot\deg_{G^{i}}(v).
Proof.

We prove the stronger statement

degGi(v)WdegHi(v)10(L+1i)degGi(v)\deg_{G^{i}}(v)\leq W\cdot\deg_{H^{i}}(v)\leq 10(L+1-i)\cdot\deg_{G^{i}}(v)

by induction from i=Li=L down to 0. For i=Li=L, since it is the last level, the entire graph GLG^{L} is a single cluster. By construction, HLH^{L} consists only of a single constant-expander H1LH^{L}_{1} that satisfies degGL(v)WdegH1L(v)9degGL(v)\deg_{G^{L}}(v)\leq W\cdot\deg_{H^{L}_{1}}(v)\leq 9\deg_{G^{L}}(v), which completes the base case of the induction.

For i<Li<L, by induction, we have WdegHi+1(v)10(Li)degGi+1(v)W\cdot\deg_{H^{i+1}}(v)\leq 10(L-i)\cdot\deg_{G^{i+1}}(v). Fix a cluster UjiU^{i}_{j} that gets contracted to vertex uji+1u^{i+1}_{j} in Gi+1G^{i+1}, and fix a vertex vUjiv\in U^{i}_{j}. For the graph H0iH^{i}_{0}, we have

degH0i(v)degHi+1(uji+1)w(EGi(v,UiUji))degGi+1(uji+1)\displaystyle\deg_{H^{i}_{0}}(v)\leq\left\lceil\deg_{H^{i+1}}(u^{i+1}_{j})\cdot\frac{w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j}))}{\deg_{G^{i+1}}(u^{i+1}_{j})}\right\rceil 1+degHi+1(uji+1)w(EGi(v,UiUji))degGi+1(uji+1)\displaystyle\leq 1+\deg_{H^{i+1}}(u^{i+1}_{j})\cdot\frac{w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j}))}{\deg_{G^{i+1}}(u^{i+1}_{j})}
1+10(Li)Ww(EGi(v,UiUji))\displaystyle\leq 1+\frac{10(L-i)}{W}w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j})) (23)
1+10(Li)WdegGi(v).\displaystyle\leq 1+\frac{10(L-i)}{W}\deg_{G^{i}}(v).

For the graph HjiH^{i}_{j}, by construction (22), we have

degGi(v)WdegHji(v)9degGi(v).\deg_{G^{i}}(v)\leq W\cdot\deg_{H^{i}_{j}}(v)\leq 9\deg_{G^{i}}(v).

Therefore,

WdegHi(v)=W(degH0i(v)+degHji(v))WdegHji(v)degGi(v)W\cdot\deg_{H^{i}}(v)=W\cdot\left(\deg_{H^{i}_{0}}(v)+\deg_{H^{i}_{j}}(v)\right)\geq W\cdot\deg_{H^{i}_{j}}(v)\geq\deg_{G^{i}}(v)

and

WdegHi(v)=W(degH0i(v)+degHji(v))W+10(Li)degGi(v)+9degGi(v).W\cdot\deg_{H^{i}}(v)=W\cdot\left(\deg_{H^{i}_{0}}(v)+\deg_{H^{i}_{j}}(v)\right)\leq W+10(L-i)\deg_{G^{i}}(v)+9\deg_{G^{i}}(v).

We can assume that Δ3\Delta\geq 3, so that degGi(v)λλ~/3=ΔW/3W\deg_{G^{i}}(v)\geq\lambda\geq\widetilde{\lambda}/3=\Delta W/3\geq W, and the above is at most

degGi(v)+10(Li)degGi(v)+9degGi(v)=10(L+1i)degGi(v),\deg_{G^{i}}(v)+10(L-i)\deg_{G^{i}}(v)+9\deg_{G^{i}}(v)=10(L+1-i)\cdot\deg_{G^{i}}(v),

which completes the induction. ∎

Claim B.4 (Analogue of (7) for HiH^{i}).

For all i,ji,j,

w(Hi[Uji]Dji)volHi[Uji](Dji)Ω(ϕβ).\frac{w(\partial_{H^{i}[U^{i}_{j}]}D^{i}_{j})}{\textbf{{vol}}_{H^{i}[U^{i}_{j}]}(D^{i}_{j})}\geq\Omega\left(\frac{\phi}{\beta}\right).
Proof.

Note that Hi[Uji]H^{i}[U^{i}_{j}] is exactly HjiH^{i}_{j} by construction. We begin by bounding volumes in GiG^{i}.

volGi(Dji)\displaystyle\textbf{{vol}}_{G^{i}}(D^{i}_{j}) =volGi[Uji](Dji)+w(EGi(Dji,UiUji))\displaystyle=\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
volGi[Uji](Dji)+βϕw(EGi(Dji,UiUji))\displaystyle\leq\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
(6)volGi(UjiDji)+βϕw(EGi(Dji,UiUji))\displaystyle\stackrel{{\scriptstyle(\ref{eq:Dij})}}{{\leq}}\textbf{{vol}}_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
βϕ(volGi(UjiDji)+w(EGi(Dji,UiUji))\displaystyle\leq\frac{\beta}{\phi}\left(\textbf{{vol}}_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})\right)
=βϕvolGi(UjiDji).\displaystyle=\frac{\beta}{\phi}\textbf{{vol}}_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j}).

We now translate this volume bound to HjiH^{i}_{j}. By the construction of HjiH^{i}_{j} (22),

volHji(Dji)9volGi(Dji)W9βvolGi(UjiDji)Wϕ9βvolHji(UjiDji)ϕ.\displaystyle\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j})\leq\frac{9\textbf{{vol}}_{G^{i}}(D^{i}_{j})}{W}\leq\frac{9\beta\textbf{{vol}}_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j})}{W\phi}\leq\frac{9\beta\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})}{\phi}. (24)

Since HjiH^{i}_{j} is an Ω(1)\Omega(1)-expander,

w(Hji(Dji))Ω(1)min{volHji(Dji),volHji(UjiDji)}Ω(1)ϕ9βvolHji(Dji),w(\partial_{H^{i}_{j}}(D^{i}_{j}))\geq\Omega(1)\cdot\min\{\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j}),\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})\}\geq\Omega(1)\cdot\frac{\phi}{9\beta}\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j}),

as desired. ∎

Claim B.5 (Analogue of (8) for HiH^{i}).

For all i<Li<L and jj,

w(Hi[Uji]Dji)w(EHi(Dji,UiUji))Ω(1L).\frac{w(\partial_{H^{i}[U^{i}_{j}]}D^{i}_{j})}{w(E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))}\geq\Omega\left(\frac{1}{L}\right).
Proof.

Let uji+1Ui+1u^{i+1}_{j}\in U^{i+1} be the vertex that the cluster UjiU^{i}_{j} contracts to in Gi+1G^{i+1}. Again, note that Hi[Uji]H^{i}[U^{i}_{j}] is exactly HjiH^{i}_{j} by construction.

The only edges of EHi(Dji,UiUji)E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}) belong to H0iH^{i}_{0}, so

EHi(Dji,UiUji)=EH0i(Dji,UiUji)=vDjidegH0i(v)\displaystyle E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})=E_{H^{i}_{0}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})=\sum_{v\in D^{i}_{j}}\deg_{H^{i}_{0}}(v) (23)vDji(1+10(Li)Ww(EGi(v,UiUji)))\displaystyle\stackrel{{\scriptstyle(\ref{eq:Hi0})}}{{\leq}}\sum_{v\in D^{i}_{j}}\left(1+\frac{10(L-i)}{W}w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j}))\right)
vDji(1+10LWw(EGi(v,UiUji)))\displaystyle\leq\sum_{v\in D^{i}_{j}}\left(1+\frac{10L}{W}w(E_{G^{i}}(v,U^{i}\setminus U^{i}_{j}))\right)
=|Dji|+10LWw(EGi(Dji,UiUji)).\displaystyle=|D^{i}_{j}|+\frac{10L}{W}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})).

We upper bound the term w(EGi(Dji,UiUji))w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})) in two ways:

βϕw(EGi(Dji,UiUji))\displaystyle\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})) volGi[Uji](Dji)+βϕw(EGi(Dji,UiUji))\displaystyle\leq\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
βϕ(volGi[Uji](Dji)+w(EGi(Dji,UiUji)))\displaystyle\leq\frac{\beta}{\phi}\left(\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\right)
=βϕvolGi(Dji);\displaystyle=\frac{\beta}{\phi}\textbf{{vol}}_{G^{i}}(D^{i}_{j});
βϕw(EGi(Dji,UiUji))\displaystyle\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})) volGi[Uji](Dji)+βϕw(EGi(Dji,UiUji))\displaystyle\leq\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
(6)volGi[Uji](UjiDji)+βϕw(EGi(Dji,UiUji))\displaystyle\stackrel{{\scriptstyle(\ref{eq:Dij})}}{{\leq}}\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(U^{i}_{j}\setminus D^{i}_{j})+\frac{\beta}{\phi}w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))
βϕ(volGi[Uji](UjiDji)+w(EGi(Dji,UiUji))\displaystyle\leq\frac{\beta}{\phi}\left(\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(U^{i}_{j}\setminus D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})\right)
=βϕvolGi(UjiDji).\displaystyle=\frac{\beta}{\phi}\textbf{{vol}}_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j}).

Therefore,

EHi(Dji,UiUji)\displaystyle E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}) |Dji|+10LWmin{volGi(Dji),volGi(UjiDji)}\displaystyle\leq|D^{i}_{j}|+\frac{10L}{W}\min\{\textbf{{vol}}_{G^{i}}(D^{i}_{j}),\textbf{{vol}}_{G^{i}}(U^{i}_{j}\setminus D^{i}_{j})\}
(22)|Dji|+10Lmin{volHji(Dji),volHji(UjiDji)}.\displaystyle\stackrel{{\scriptstyle(\ref{eq:Hj})}}{{\leq}}|D^{i}_{j}|+10L\min\{\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j}),\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})\}\ .

We now bound |Dji||D^{i}_{j}| as follows. By construction (22), for all vUjiv\in U^{i}_{j},

degHji(v)degGi(v)WλWλ~3W=Δ3,\deg_{H^{i}_{j}}(v)\geq\frac{\deg_{G^{i}}(v)}{W}\geq\frac{\lambda}{W}\geq\frac{\widetilde{\lambda}}{3W}=\frac{\Delta}{3},

which means that

|Dji|volHji(Dji)Δ/3(24)27βvolHji(UjiDji)ΔϕvolHji(UjiDji)|D^{i}_{j}|\leq\frac{\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j})}{\Delta/3}\stackrel{{\scriptstyle(\ref{eq:vol})}}{{\leq}}\frac{27\beta\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})}{\Delta\phi}\leq\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})

as long as we impose the condition

Δ27βϕ.\displaystyle\Delta\geq\frac{27\beta}{\phi}. (25)

Therefore,

|Dji|min{volHji(Dji),volHji(UjiDji)}|D^{i}_{j}|\leq\min\{\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j}),\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})\}

and

w(EHi(Dji,UiUji))(10L+1)min{volHji(Dji),volHji(UjiDji)}.w(E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\leq(10L+1)\min\{\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j}),\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})\}.

Since HjiH^{i}_{j} is an Ω(1)\Omega(1)-expander,

w(HjiDji)Ω(1)min{volHji(Dji),volHji(UjiDji)}Ω(1L)w(EHi(Dji,UiUji)),w(\partial_{H^{i}_{j}}D^{i}_{j})\geq\Omega(1)\cdot\min\{\textbf{{vol}}_{H^{i}_{j}}(D^{i}_{j}),\textbf{{vol}}_{H^{i}_{j}}(U^{i}_{j}\setminus D^{i}_{j})\}\geq\Omega\left(\frac{1}{L}\right)\cdot w(E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})),

as desired. ∎

Lemma B.6.

For all i,ji,j,

Ω(ϕβ)w(GiDji)Ww(HiDji)O(Lϕ)w(GiDji).\Omega\left(\frac{\phi}{\beta}\right)w(\partial_{G^{i}}D^{i}_{j})\leq W\cdot w(\partial_{H^{i}}D^{i}_{j})\leq O\left(\frac{L}{\phi}\right)w(\partial_{G^{i}}D^{i}_{j}).
Proof.

For the lower bound, we have

w(HiDji)w(Hi[Uji]Dji)\displaystyle w(\partial_{H^{i}}D^{i}_{j})\geq w(\partial_{H^{i}[U^{i}_{j}]}D^{i}_{j})\quad Clm.B.4Ω(ϕβ)volHi[Uji](Dji)\displaystyle\stackrel{{\scriptstyle\mathclap{\text{Clm.}\ref{clm:Exp}}}}{{\geq}}\quad\Omega\left(\frac{\phi}{\beta}\right)\textbf{{vol}}_{H^{i}[U^{i}_{j}]}(D^{i}_{j})
(22)Ω(ϕβ)1WvolGi(Dji)\displaystyle\stackrel{{\scriptstyle\mathclap{(\ref{eq:Hj})}}}{{\geq}}\Omega\left(\frac{\phi}{\beta}\right)\cdot\frac{1}{W}\textbf{{vol}}_{G^{i}}(D^{i}_{j})
Ω(ϕβ)1Ww(GiDji).\displaystyle\geq\Omega\left(\frac{\phi}{\beta}\right)\cdot\frac{1}{W}w(\partial_{G^{i}}D^{i}_{j}).

For the upper bound,

w(HiDji)=w(Hi[Uji]Dji)+w(EHi(Dji,UiUji))\displaystyle w(\partial_{H^{i}}D^{i}_{j})=w(\partial_{H^{i}[U^{i}_{j}]}D^{i}_{j})+w(E_{H^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j})) Clm.B.5(1+O(L))w(Hi[Uji]Dji)\displaystyle\stackrel{{\scriptstyle\mathclap{\text{Clm.}\ref{clm:BL}}}}{{\leq}}(1+O(L))\cdot w(\partial_{H^{i}[U^{i}_{j}]}D^{i}_{j})
(1+O(L))volHi[Uji](Dji)\displaystyle\leq(1+O(L))\cdot\textbf{{vol}}_{H^{i}[U^{i}_{j}]}(D^{i}_{j})
(22)(1+O(L))9WvolGi(Dji)\displaystyle\stackrel{{\scriptstyle\mathclap{(\ref{eq:Hj})}}}{{\leq}}(1+O(L))\cdot\frac{9}{W}\textbf{{vol}}_{G^{i}}(D^{i}_{j})
(1+O(L))9W(volGi[Uji](Dji)+w(EGi(Dji,UiUji)))\displaystyle\leq(1+O(L))\cdot\frac{9}{W}\left(\textbf{{vol}}_{G^{i}[U^{i}_{j}]}(D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\right)
(1+O(L))9W(1ϕw(Gi[Uji]Dji)+w(EGi(Dji,UiUji)))\displaystyle\leq(1+O(L))\cdot\frac{9}{W}\left(\frac{1}{\phi}w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\right)
(1+O(L))9W1ϕ(w(Gi[Uji]Dji)+w(EGi(Dji,UiUji)))\displaystyle\leq(1+O(L))\cdot\frac{9}{W}\cdot\frac{1}{\phi}\left(w(\partial_{G^{i}[U^{i}_{j}]}D^{i}_{j})+w(E_{G^{i}}(D^{i}_{j},U^{i}\setminus U^{i}_{j}))\right)
=(1+O(L))9W1ϕw(GiDji).\displaystyle=(1+O(L))\cdot\frac{9}{W}\cdot\frac{1}{\phi}w(\partial_{G^{i}}D^{i}_{j}).

Lemma B.7.

WHW\cdot H is a γ\gamma-approximate sparsifier with γ=max{O(βO(L)L/ϕ),O(LO(L)β/ϕ)}\gamma=\max\left\{O(\beta^{-O(L)}L/\phi),O(L^{O(L)}\beta/\phi)\right\}.

Proof.

Since Claim B.5 is an analogue of (8) for graph HH with the parameter β\beta replaced by Ω(1/L)\Omega(1/L), we can apply Lemmas 3.3 and 3.4 to HH, obtaining

w(HS)i=0Lj[ki]w(HiDji)LO(L)w(HS).\displaystyle w(\partial_{H}S)\leq\sum_{i=0}^{L}\sum_{j\in[k_{i}]}w(\partial_{H^{i}}D^{i}_{j})\leq L^{O(L)}w(\partial_{H}S).

Combining this with Lemma B.6,

w(HS)i=0Lj[ki]w(HiDji)\displaystyle w(\partial_{H}S)\leq\sum_{i=0}^{L}\sum_{j\in[k_{i}]}w(\partial_{H^{i}}D^{i}_{j}) (1+O(L))9W1ϕi=0Lj[ki]w(GiDji)\displaystyle\leq(1+O(L))\cdot\frac{9}{W}\cdot\frac{1}{\phi}\sum_{i=0}^{L}\sum_{j\in[k_{i}]}w(\partial_{G^{i}}D^{i}_{j})
Lem.3.4(1+O(L))9W1ϕβO(L)w(GS)\displaystyle\stackrel{{\scriptstyle\mathclap{\text{Lem.}\ref{lem:D-ub}}}}{{\leq}}\quad(1+O(L))\cdot\frac{9}{W}\cdot\frac{1}{\phi}\cdot\beta^{-O(L)}w(\partial_{G}S)

and

w(HS)LO(L)i=0Lj[ki]w(HiDji)\displaystyle w(\partial_{H}S)\geq L^{-O(L)}\sum_{i=0}^{L}\sum_{j\in[k_{i}]}w(\partial_{H^{i}}D^{i}_{j}) LO(L)i=0Lj[ki]Ω(ϕβ)1Ww(GiDji)\displaystyle\geq L^{-O(L)}\cdot\sum_{i=0}^{L}\sum_{j\in[k_{i}]}\Omega\left(\frac{\phi}{\beta}\right)\frac{1}{W}w(\partial_{G^{i}}D^{i}_{j})
Lem.3.3LO(L)Ω(ϕβ)1Ww(GS).\displaystyle\stackrel{{\scriptstyle\mathclap{\text{Lem.}\ref{lem:D-lb}}}}{{\geq}}\quad L^{-O(L)}\cdot\Omega\left(\frac{\phi}{\beta}\right)\frac{1}{W}w(\partial_{G}S).

Finally, we set the parameters r1,β,L,ϕr\geq 1,\beta,L,\phi. For r:=(logn)1/6r:=(\log n)^{1/6}, let β:=(logn)O(r4)\beta:=(\log n)^{-O(r^{4})} and ϕ:=(logn)r5\phi:=(\log n)^{-r^{5}}, so that by Theorem 3.2, the total weight of inter-cluster edges, and therefore the total weight of the next graph in the expander decomposition sequence, shrinks by factor (logn)O(r4)ϕ=(logn)Ω(r5)(\log n)^{O(r^{4})}\phi=(\log n)^{-\Omega(r^{5})}. Since edge weights are assumed to be polynomially bounded, this shrinking can only happen O(lognr5)O(\frac{\log n}{r^{5}}) times, so LO(lognr5)L\leq O(\frac{\log n}{r^{5}}). Therefore, our approximation factor is

γ=max{O(βO(L)L/ϕ),O(LO(L)β/ϕ)}=O(βO(L)L/ϕ)=2O(logn)5/6(loglogn)O(1),\gamma=\max\left\{O(\beta^{-O(L)}L/\phi),O(L^{O(L)}\beta/\phi)\right\}=O(\beta^{-O(L)}L/\phi)=2^{O(\log n)^{5/6}(\log\log n)^{O(1)}},

and the running time, which is dominated by the output size O(Δm)O(\Delta m) and the calls to Theorem 3.2 and Lemma A.4, is

O(Δm)+m1+O(1/r)+O~(m/ϕ2)=2O(logn)5/6(loglogn)O(1)m+O(Δm).O(\Delta m)+m^{1+O(1/r)}+\widetilde{O}(m/\phi^{2})=2^{O(\log n)^{5/6}(\log\log n)^{O(1)}}m+O(\Delta m).

Finally, the condition Δ27βϕ\Delta\geq\frac{27\beta}{\phi} from (25) becomes Δ2Ω(logn)5/6\Delta\geq 2^{\Omega(\log n)^{5/6}}, concluding the proof of Theorem 3.14.