This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Differentially Private Gomory-Hu Trees

Anders Aamand BARC, University of Copenhagen. aamand@mit.edu. Supported by VILLUM Foundation grant 16582 and DFF-International Postdoc Grant 0164-00022B from the Independent Research Fund Denmark.    Justin Y. Chen Massachusetts Institute of Technology. justc@mit.edu. Supported by an NSF Graduate Research Fellowship under Grant No. 17453. Part of this work was conducted while the author was visiting the Simons Institute for the Theory of Computing.    Mina Dalirrooyfard Morgan Stanley. minad@mit.edu.    Slobodan Mitrović UC Davis. smitrovic@ucdavis.edu. Supported by the Google Research Scholar and NSF Faculty Early Career Development Program No. 2340048. Part of this work was conducted while the author was visiting the Simons Institute for the Theory of Computing.    Yuriy Nevmyvaka Morgan Stanley. yuriy.nevmyvaka@morganstanley.com.    Sandeep Silwal UW-Madison. silwal@cs.wisc.edu.    Yinzhan Xu Massachusetts Institute of Technology. xyzhan@mit.edu. Supported by NSF Grant CCF-2330048 and a Simons Investigator Award.
Abstract

Given an undirected, weighted nn-vertex graph G=(V,E,w)G=(V,E,w), a Gomory-Hu tree TT is a weighted tree on VV such that for any pair of distinct vertices s,tVs,t\in V, the Min-ss-tt-Cut on TT is also a Min-ss-tt-Cut on GG. Computing a Gomory-Hu tree is a well-studied problem in graph algorithms and has received considerable attention. In particular, a long line of work recently culminated in constructing a Gomory-Hu tree in almost linear time [Abboud, Li, Panigrahi and Saranurak, FOCS 2023].

We design a differentially private (DP) algorithm that computes an approximate Gomory-Hu tree. Our algorithm is ε\varepsilon-DP, runs in polynomial time, and can be used to compute ss-tt cuts that are O~(n/ε)\tilde{O}(n/\varepsilon)-additive approximations of the Min-ss-tt-Cuts in GG for all distinct s,tVs,t\in V with high probability. Our error bound is essentially optimal, as [Dalirrooyfard, Mitrović and Nevmyvaka, NeurIPS 2023] showed that privately outputting a single Min-ss-tt-Cut requires Ω(n)\Omega(n) additive error even with (1,0.1)(1,0.1)-DP and allowing for a multiplicative error term. Prior to our work, the best additive error bounds for approximate all-pairs Min-ss-tt-Cuts were O(n3/2/ε)O(n^{3/2}/\varepsilon) for ε\varepsilon-DP [Gupta, Roth and Ullman, TCC 2012] and O(mnpolylog(n/δ)/ε)O(\sqrt{mn}\cdot\text{polylog}(n/\delta)/\varepsilon) for (ε,δ)(\varepsilon,\delta)-DP [Liu, Upadhyay and Zou, SODA 2024], both of which are implied by differential private algorithms that preserve all cuts in the graph. An important technical ingredient of our main result is an ε\varepsilon-DP algorithm for computing minimum Isolating Cuts with O~(n/ε)\tilde{O}(n/\varepsilon) additive error, which may be of independent interest.

1 Introduction

Given an undirected, weighted graph G=(V,E,w)G=(V,E,w) with positive edge weights, a cut is a bipartition of vertices (U,VU)(U,V\setminus U), and the value of the cut is the total weights of edges crossing the bipartition. Given a pair of distinct vertices s,tVs,t\in V, the Min-ss-tt-Cut is a minimum-valued cut where sUs\in U and tVUt\in V\setminus U. Min-ss-tt-Cut is dual to the Max-ss-tt-Flow problem, and the celebrated max-flow min-cut theorem states that the value of the Min-ss-tt-Cut equals the value of the Max-ss-tt-Flow [FF56, EFS56]. Finding a Min-ss-tt-Cut (or equivalently Max-ss-tt-Flow) is a fundamental problem in algorithmic graph theory, which has been studied for over seven decades [Dan51, HR55, FF56, EFS56], and has inspired ample algorithmic research (e.g., see a survey in [CKL+22, Appendix A]). It also has a wide range of algorithmic applications, including edge connectivity [Men27], bipartite matching (see, e.g., [CLRS22]), minimum Steiner cut [LP20], vertex-connectivity oracles [PSY22], among many others.

A natural and well-studied variant of Min-ss-tt-Cut is the All-Pairs Min Cut (APMC) problem. Given an input graph, the goal is to output the Min-ss-tt-Cut for all the pairs of vertices ss and tt in VV. In a seminal paper, Gomory and Hu [GH61] showed that there is a tree representation for this problem, called GH-tree or cut tree, that takes only n1n-1 Min-ss-tt-Cut (Max Flow) calls to compute. Consequently, there are only n1n-1 different max flow / minimum cut values in an arbitrary graph with positive edge weights. There has been a long line of research in designing faster GH-tree algorithms (e.g. [HKPB07, BSWN15, AKT20, Zha21, LP21, AKT21, AKT22, ALPS23], also see the survey [Pan08]), culminating an almost linear time algorithm for computing the GH-tree [ALPS23]. Beyond theoretical considerations, the GH-Tree has a long list of applications in a variety of research areas, such as networks [Hu74], image processing [WL93], optimization such as the bb-matching problem [PR82] and webpage segmenting [LLT11]. Furthermore, the global Min Cut of a graph is easily obtainable from the GH-tree.

In practice, algorithms are often applied to large data sets containing sensitive information. It is well understood by now that even minor negligence in handling users’ data can have severe consequences on user privacy; see [BDK07, NS08, Kor10, SSSS17, CRB+19] for a few examples. Differential privacy (DP), introduced by Dwork, McSherry, Nissim, and Smith in their seminal work [DMNS06], is a widely adopted standard for formalizing the privacy guarantees of algorithms. Informally, an algorithm is deemed differentially private if the outputs for two given neighboring inputs are statistically indistinguishable. The notion of neighboring inputs is application-dependent. In particular, when the underlying input belongs to a family of graphs, arguably the most studied notion of DP is with respect to edges, in which two neighboring graphs differ in only one edge; if the graphs are weighted, two neighboring graphs are those whose weights differ by at most one and in a single edge111Algorithms satisfying this notion are often also private for graphs whose vector of (n2)\binom{n}{2} edge weights differ by 11 in 1\ell_{1} distance.. This is the setting for all the of the private cut problems we discuss in this work. In a related notion called vertex DP, two graphs are called neighboring if they differ in the existence of one vertex and the edges incident to that vertex. Two qualitative notions of DP that formalize the meaning of being “statistically indistinguishable” are pure and approximate DP. In pure DP, which is the focus of this paper, privacy is measured by a parameter ε\varepsilon, and we use ε\varepsilon-DP to refer to an algorithm having pure DP. In approximate DP, the privacy is expressed using two parameters, ε\varepsilon and δ\delta where an individual’s data can be leaked with a small probability corresponding to δ\delta. Pure DP implies approximate DP and is a qualitatively stronger privacy guarantee. We formally define these notions in Section 2.

Over the last two decades, the design of DP graph algorithms has gained significant attention [MT07, HLMJ09, RHMS09, GLM+10, KRSY11, GRU12, BBDS13, KNRS13, BLR13, BNSV15, Sea16, AU19, US19, HKM+20, EKKL20, BEK21, RSSS21, NV21, BKM+22, FHS22, DLR+22, FLL22, CGK+23, DLL23, DKLV23, ELRS23, WRRW23, LUZ24b, JKR+24]. Several of those works focus on approximating graph cuts in the context of DP. For instance, Gupta et al. [GLM+10] gave an ε\varepsilon-DP algorithm for global Min Cut with additive error O(logn/ε)O(\log{n}/\varepsilon). The authors also showed that there does not exist an ε\varepsilon-DP algorithm for global Min Cut incurring less than Ω(logn)\Omega{(\log{n})} additive error.

Gupta, Roth, and Ullman [GRU12] and, independently, Blocki Blum, Datta, and Sheffet [BBDS12] focused on preserving all graph cuts in a DP manner. Given a graph GG, their algorithms output a synthetic graph HH such that each cut-value in HH is the same as the corresponding cut-value in GG up to an additive error of O(n1.5/ε)O\left(n^{1.5}/\varepsilon\right). The former result is pure DP while the latter is approximate DP. Eliáš, Kapralov, Kulkarni and Lee [EKKL20] improved on these results for sparse, unweighted (or small total weight) graphs, achieving error O~(mn/ε)\tilde{O}\left(\sqrt{mn/\varepsilon}\right)222In this work, the notation O~(x)\tilde{O}(x) stands for O(xpolylogx)O(x\cdot\operatorname{polylog}x). with approximate DP. The authors show that this error is essentially tight. In a follow-up work, Liu, Upadhyay, and Zou [LUZ24b] extended these results to weighted graphs and gave an algorithm to release a synthetic graph with O~(mn/ε)\tilde{O}\left(\sqrt{mn}/\varepsilon\right) error. A recent paper [LUZ24a] gives an algorithm for releasing a synthetic graph with worse error O~(m/ε)\tilde{O}(m/\varepsilon) but which runs in near-linear time.

For a different problem, Dalirrooyfard, Mitrović and Nevmyvaka [DMN23] gave an ε\varepsilon-DP algorithm for the Min-ss-tt-Cut problem with additive error O(n/ε)O(n/\varepsilon) and showed an essentially matching Ω(n)\Omega(n) lower bound, even with approximate DP and both multiplicative and additive error.

Since the algorithm by [LUZ24b] approximately preserves all the cuts, it can also be used to solve APMC in a DP manner, albeit with approximate DP and with the additive error of O~(mn/ε)\tilde{O}\left(\sqrt{mn}/\varepsilon\right). Additionally, the Ω(n)\Omega(n) lower bound for Min-ss-tt-Cut shown by [DMN23] also applies to APMC, as the latter is a harder problem. While the additive error for computing global Min Cut [GLM+10], Min-ss-tt-Cut [DMN23], and all-cuts [EKKL20, LUZ24b] while ensuring DP is tightly characterized up to polylogn\operatorname{polylog}n and 1/ε1/\varepsilon factors, there still remains a gap of roughly m/n\sqrt{m/n} between the best known lower and upper bound for solving DP APMC.333We remark that for the mentioned problems there still remain several interesting questions. For instance, finding a pure DP global Min Cut with O(logn/ε)O(\log n/\varepsilon) additive error in polynomial time is unknown. The pure DP algorithm presented in [GLM+10] runs in exponential time, and can be improved to polynomial time only at the expense of approximate DP. This motivates the following question:

Can we obtain tight bounds on the additive error for APMC with differential privacy?

1.1 Our Contribution

Problem Additive Error DP Output Runtime
Global Min Cut [GLM+10] Θ(log(n)/ε)\Theta(\log(n)/\varepsilon) \cellcolor[HTML]9AFF99Pure \cellcolor[HTML]9AFF99Cut \cellcolor[HTML]FFCCC9Exponential
Global Min Cut [GLM+10] Θ(log(n)/ε)\Theta(\log(n)/\varepsilon) \cellcolor[HTML]FFFC9EApprox \cellcolor[HTML]9AFF99Cut \cellcolor[HTML]9AFF99Polynomial
Min-ss-tt-Cut [DMN23] O(n/ε)O(n/\varepsilon) and Ω(n)\Omega(n) \cellcolor[HTML]9AFF99 Pure \cellcolor[HTML]9AFF99Cut \cellcolor[HTML]9AFF99Polynomial
All Cuts [GRU12] O(n3/2/ε)O(n^{3/2}/\varepsilon) \cellcolor[HTML]9AFF99Pure \cellcolor[HTML]9AFF99Synthetic Graph \cellcolor[HTML]9AFF99Polynomial
All Cuts  [EKKL20, LUZ24b] O~(mn/ε)\tilde{O}(\sqrt{mn}/\varepsilon) and Ω(mn/ε)\Omega(\sqrt{mn/\varepsilon}) \cellcolor[HTML]FFFC9EApprox \cellcolor[HTML]9AFF99Synthetic Graph \cellcolor[HTML]9AFF99Polynomial
APMC Values (Trivial) O(n/ε)O(n/\varepsilon) \cellcolor[HTML]FFFC9EApprox \cellcolor[HTML]FFFC9EValues Only \cellcolor[HTML]9AFF99Polynomial
APMC (Our Work)
O~(n/ε)\tilde{O}(n/\varepsilon) \cellcolor[HTML]9AFF99Pure \cellcolor[HTML]9AFF99GH-Tree \cellcolor[HTML]9AFF99Polynomial
Table 1: State-of-the-art bounds for various cut problems with differential privacy. Dependencies on the approximate DP parameter δ\delta are hidden. The APMC values result with approximate DP follows from advanced composition by adding Lap(O(n/ε))(O(n/\varepsilon)) random noise to all (n2)\binom{n}{2} true values. For APMC, the lower bound of Ω(n)\Omega(n) error [DMN23] also applies as APMC generalizes Min-ss-tt-Cut.

Our main contribution is an ε\varepsilon-DP algorithm for APMC with O~(n/ε)\tilde{O}(n/\varepsilon) additive error. Up to polylog(n)\operatorname{polylog}(n) factors, our algorithm privately outputs all the Min-ss-tt-Cuts while incurring the same error required to output an Min-ss-tt-Cut for a single pair of vertices ss and tt. This closes the aforementioned gap in the literature in private cut problems and shows that the error required to privatize the APMC problem is closer to that of the Min-ss-tt-Cut problem than to the synthetic graph/all-cuts problem.

To achieve this result, we introduce a DP algorithm that outputs an approximate Gomory-Hu tree (GH-tree). Gomory and Hu [GH61] showed that for any undirected graph GG, there exists a tree TT defined on the vertices of graph GG such that for all pairs of vertices s,ts,t, the Min-ss-tt-Cut in TT is also a Min-ss-tt-Cut in GG. We develop a private algorithm for constructing such a tree.

Theorem 1.1.

Given a weighted graph G=(V,E,w)G=(V,E,w) with positive edge weights and a privacy parameter ε>0\varepsilon>0, there exists an ε\varepsilon-DP algorithm that outputs an approximate GH-tree TT with additive error O~(n/ε)\tilde{O}(n/\varepsilon): for any stVs\neq t\in V, the Min-ss-tt-Cut on TT and the Min-ss-tt-Cut on GG differ in O~(n/ε)\tilde{O}(n/\varepsilon) in cut value with respect to edge weights in GG. The algorithm runs in time O~(n2)\tilde{O}(n^{2}), and the additive error guarantee holds with high probability.

Since the GH-tree output by Theorem 1.1 is private, any post-processing on this tree is also private. This yields the following corollary.

Corollary 1.1.

Given a weighted graph GG with positive edge weights and a privacy parameter ε>0\varepsilon>0, there exists an ε\varepsilon-DP algorithm that outputs a cut for all the pairs of vertices ss and tt whose value is within O~(n/ε)\tilde{O}(n/\varepsilon) from the value of the Min-ss-tt-Cut with high probability.

Note that Corollary 1.1 is tight up to polylog(n)\operatorname{polylog}(n) and 1ϵ\frac{1}{\epsilon} factors since any (pure or approximate) DP algorithm outputting the Min-ss-tt-Cut for fixed pair of vertices ss and tt requires Ω(n)\Omega{(n)} additive error [DMN23]. Another corollary of Theorem 1.1 is a polynomial time pure DP algorithm for the global Min-Cut problem.

Corollary 1.2.

Given a weighted graph GG with positive edge weights and a privacy parameter ε>0\varepsilon>0, there exists an ε\varepsilon-DP algorithm that outputs an approximate global Min-Cut of GG in O~(n2)\tilde{O}(n^{2}) time and has additive error O~(n/ε)\tilde{O}(n/\varepsilon) with high probability.

Note that [GLM+10] obtained an exponential time pure DP algorithm and a polynomial time approximate DP algorithm for global Min-Cut. It remains an intriguing question whether pure DP global Min-Cut can be found with only polylogn/ε\operatorname{polylog}n/\varepsilon additive error and in polynomial time.

Lastly, we note an application to the minimum kk-cut problem. Here, the goal is to partition the vertex set into kk pieces and the cost of any partitioning is weight of all edges that cross partitions. We wish to find the smallest cost solution. It is known that simply removing the smallest k1k-1 edges of an exact GH tree gives us a solution to the minimum kk-cut problem with a multiplicative approximation of 22 [SV95]. Since we compute an approximate GH tree with additive error O~(n/ε)\tilde{O}(n/\varepsilon), we can obtain a solution to minimum kk-cut with multiplicative error 22 and additive error O~(nk/ε)\tilde{O}(nk/\varepsilon). We give the proof of the following corollary in Appendix A.

Corollary 1.3.

Given a weighted graph GG with positive edge weights and a privacy parameter ε>0\varepsilon>0, there exists an ε\varepsilon-DP algorithm that outputs a solution to the minimum kk-cut problem on GG in O~(n2)\tilde{O}(n^{2}) time with multiplicative error 22 and additive error O~(nk/ε)\tilde{O}(nk/\varepsilon) with high probability.

The only prior DP algorithm for the minimum kk-cut problem is given in [CDFZ24]. It requires approximate DP and has the same multiplicative error 22 but additive error O~(k1.5/ε)\tilde{O}(k^{1.5}/\varepsilon). [CDFZ24] also give a pure DP algorithm which only handles unweighted graphs and requires exponential time. To the best of our knowledge, there are no prior pure DP algorithms which can compute the minimum kk-cut on weighted graphs. It is an interesting question to determine the limits of pure DP and efficient (polynomial time) algorithms for the minimum kk-cut problem.

1.2 Technical Overview

A greatly simplified view of a typical approach to designing a DP algorithm is to start with its non-DP version and then privatize it. The main challenge is finding the right way to privatize an algorithm, assuming a way exists in the first place, and then proving that DP guarantees are indeed achieved. To provide a few examples, Gupta et al. [GLM+10] employ Karger’s algorithm [Kar93] to produce a set of cuts and then use the Exponential Mechanism [MT07] to choose one of those cut. This simple but clever approach results in an (ε,δ)(\varepsilon,\delta)-DP algorithm for global Min Cut, for δ=O(1/poly(n))\delta=O(1/\operatorname{poly}(n)), with O(logn/ε)O(\log{n}/\varepsilon) additive error. Dalirrooyfard et al. [DMN23] show that the following simple algorithm yields ε\varepsilon-DP Min-ss-tt-Cut with O(n/ε)O(n/\varepsilon) additive error: for each vertex vv, add an edge from vv to ss and from vv to tt with their weights chosen from the exponential distribution with parameter 1/ε1/\varepsilon; return the Min-ss-tt-Cut on the modified graph. In the rest of this section, we first describe why directly privatizing some of the existing non-private algorithms does not yield the advertised additive error. Then, we describe our approach.

1.2.1 Obstacles in privatizing Gomory-Hu Trees

The All-Pairs Min-Cut (APMC) problem outputs cuts for Θ(n2)\Theta(n^{2}) pairs but there are known to be only O(n)O(n) distinct cuts. This property was leveraged in the pioneering work by Gomory and Hu [GH61], who introduced the Gomory-Hu tree (GH-tree), a structure that succinctly represents all the (n2)\binom{n}{2} Min-ss-tt-Cuts in a graph.

A Gomory-Hu tree is constructed through a recursive algorithm that solves the Min-ss-tt-Cut problem at each recursion step. The algorithm can be outlined as follows: (1) Start with one supernode in a tree TT, consisting of all the vertices of the input graph GG. (2) At each step, select an arbitrary supernode SS where |S|>1|S|>1 and choose two arbitrary vertices ss and tt within it. (3) Consider the graph where all supernodes, except SS, are each contracted into a single vertex. Compute the Min-ss-tt-Cut in this contracted graph. (4) Using the result of this Min-ss-tt-Cut and the submodularity of cuts (see Lemma 2.1), update the tree TT to ensure ss and tt are in different supernodes. The GH-tree efficiently represents all pairwise min-cuts in the graph by iterating through these steps.

One can replace the Min-ss-tt-Cut procedures with the private Min-ss-tt-Cut algorithm of [DMN23] to privatize this algorithm. Several obstacles arise when attempting to turn this idea into an error-efficient algorithm. Firstly, the algorithm of [DMN23] changes the graph. It is unclear whether we need to keep these changes for each run of the Min-ss-tt-Cut or revert back to the original graph. Secondly, the depth of this recursion could be O(n)O(n), which means that using Basic Composition [DKM+06], we obtain an εn\varepsilon n-DP algorithm with additive error O(n2/ε)O(n^{2}/\varepsilon). Even if we apply Advanced Composition [DR14], which will result in an approximate DP algorithm, the resulting algorithm would be (εn,δ)(\varepsilon\sqrt{n},\delta)-DP with additive error of O(n2log(1/δ)/ε)O(n^{2}\log(1/\delta)/\varepsilon). This is significantly worse then prior work preserving all cuts.

1.2.2 DP Approximate Min Isolating Cuts (Section 3)

The above approach suggests attempting to privatize a GH-tree algorithm with low recursion depth. There has been recent interest in low-depth recursion GH-tree algorithms (e.g., [AKT21, AKL+22, LPS22, ALPS23]), intending to construct GH-tree using fewer than O(n)O(n) Min-ss-tt-Cut, i.e., Max-Flow, computations. Most of these works use an important sub-routine, called Min Isolating Cuts introduced by [LP20, AKT21]. For a set UU of vertices in GG, let w(U)w(U) be the sum of the weights of the edges in GG with exactly one endpoint in UU.

Definition 1.1 (Min Isolating Cuts [LP20, AKT21]).

Given a set of terminals RVR\subseteq V, the Min Isolating Cuts problem asks to output a collection of sets {SvV:vR}\{S_{v}\subseteq V:v\in R\} such that for each vertex vRv\in R, the set SvS_{v} satisfies SvR={v}S_{v}\cap R=\{v\}, and it has the minimum value of w(Sv)w(S^{\prime}_{v}) over all sets SvS^{\prime}_{v} that SvR={v}S^{\prime}_{v}\cap R=\{v\}. In other words, each SvS_{v} is the minimum cut separating vv from R{v}R\setminus\{v\}.

[LP20] and [AKT21] independently introduce the Isolating Cuts Lemma, showing how to solve the Min Isolating Cuts problem in O(logR)O(\log{R}) many Min-ss-tt-Cut runs. [LP20] use it to find the global Min Cut in poly logarithmically many invocations of Max Flow and [AKT21] use it to compute GH-tree in simple graphs. Subsequent algorithms for GH-tree also use the Isolating Cuts Lemma, including the almost linear time algorithm for weighted graphs [ALPS23]. The Isolating Cuts Lemma has been extended to obtain new algorithms for finding the non-trivial minimizer of a symmetric submodular function and solving the hypergraph minimum cut problem [MN21, CQ21]. We develop the first differentially private algorithm for finding Min Isolating Cuts.

Theorem 1.2.

There is an ε\varepsilon-DP algorithm that given a graph GG and a set of terminals RR, outputs sets {Sv:vR}\{S_{v}:v\in R\}, such that for each vertex vRv\in R, the set SvS_{v} satisfies SvR={v}S_{v}\cap R=\{v\}, and w(Sv)w(Sv)+O~(n/ε)w(S_{v})\leq w(S^{*}_{v})+\tilde{O}(n/\varepsilon) with high probability, where {Sv:vR}\{S^{*}_{v}:v\in R\} are the Min Isolating Cuts for RR.

Our overall algorithm builds on a reduction first shown by [AKT20] from GH-tree to single-source Min Cuts, where we need to compute the ss-tt min cut values for a fixed ss and every tVt\in V. Recent almost linear time algorithms for GH-tree [AKL+22, ALPS23] also utilize this reduction. We will mostly follow an exposition by Li [Li21].

We now provide a high-level overview of the algorithm of [Li21]. The approach of [Li21] is a recursive algorithm with depth polylog(n)\operatorname{polylog}(n). A generalized version of GH-tree called Steiner GH-Tree was considered, where instead of looking for Min-ss-tt-Cut for all pairs of vertices ss and tt in GG, we only focus on pairs s,tUs,t\in U where UU is a subset of vertices. This especially helps in recursive algorithms, preventing the algorithm from computing the Min-ss-tt-Cut of the pairs of vertices s,ts,t that appear in multiple instances of the problem multiple times.

Each recursive step receives a set of terminals UU, a source terminal sUs\in U, and a graph GG. The goal of each recursive step is to decompose the graph into subgraphs of moderate sizes, enabling further recursion on each subgraph. To achieve this, the algorithm utilizes the Min Isolating Cuts Lemma. Specifically, it applies the Min Isolating Cuts Lemma to random subsets UUU^{\prime}\subseteq U. For each UU^{\prime} and vUv\in U^{\prime}, a Min Isolating Cut SvS_{v} is computed, which separates vv from all terminals in UU^{\prime}. Note that vertices in UUU\setminus U^{\prime} may be included in SvS_{v}. The algorithm considers a set SvS_{v} to be “good” if w(Sv)w(S_{v}) equals the Min-ss-vv-Cut and SvS_{v} contains at most |U|/2|U|/2 terminals. This criterion ensures that, during the recursion, the number of terminals decreases exponentially, resulting in the recursion depth being polylog(n)\operatorname{polylog}(n).

[Li21] shows that in one recursive step, one can obtain a collection of disjoint “good” sets SvS_{v}. Let SlargeS_{\text{large}} be the set of vertices that are not in any SvS_{v}. The authors show that SlargeS_{\text{large}} has at most |U|(11/polylog(n))|U|\cdot(1-1/\operatorname{polylog}(n)) many terminals in UU. The recursion branches out as follows: for each SvS_{v}, contract all the vertices that are not in SvS_{v}, and recurse on this graph GvG_{v}. For SlargeS_{\text{large}}, contract each SvS_{v} into a vertex, and recurse on this graph GlargeG_{\text{large}}. It is similar to the original GH-tree algorithm [GH61] to aggregate the output of these recursions into a GH-tree.444Much of the effort in the recent works on constructing GH-tree is directed on obtaining Min-ss-vv-Cut values from ss to all vv in the recursive step efficiently. From the point of view of privacy and error, this step is trivial as single source minimum cuts can be released with O(n/ε)O(n/\varepsilon) error via basic composition as there are nn values.

1.2.3 DP Gomory-Hu Tree (Sections 4 and 5)

We now outline the main challenges and how we overcome them in obtaining a differentially private GH-tree.

Approximately good sets.

We will follow the high-level strategy of using a Min Isolating Cuts algorithm (now via our DP algorithm described above) to find sets SvS_{v} which are approximately Min-ss-tt-Cuts. In defining which approximate Min Isolating Cuts are “good” and which our algorithm will recurse on, we want to accomplish three goals:

  1. 1.

    The sets SvS_{v} we return are O~(n/ε)\tilde{O}(n/\varepsilon)-additive approximate Min-ss-vv-Cuts.

  2. 2.

    Each set SvS_{v} does not contain more than a constant fraction of UU in order to bound the recursion depth on SvS_{v}.

  3. 3.

    The union of sets SvS_{v} we return contain a 1/polylog(n)1/\operatorname{polylog}(n) fraction of the vertices in UU, this ensures that recursing on SlargeS_{\text{large}} has reasonably bounded depth.

The first goal can be achieved by privately estimating the value of each Min-ss-vv-Cut and ensuring that w(Sv)w(S_{v}) is close to this value up to approximation errors due to privacy.

To address the second goal, we only require a good set SvS_{v} to have at most 0.9|U|0.9|U| terminals in UU. To bias the algorithm towards selecting smaller sets, we use the following idea. Consider a graph GG with terminals ss and tt, and a subset of vertices UU. We aim to obtain a DP Min-ss-tt-Cut such that the ss-side of the cut contains a small number of vertices in UU, without significantly sacrificing accuracy. To achieve this, we add edges from tt to every vertex in UU with a certain weight, penalizing the placement of vertices in UU on the ss-side of the cut. By increasing the error of our Min Isolating Cuts algorithm by only a constant factor, we enforce that if a true minimum isolating cut of size |U|/2|U|/2 exists, we will output an isolating of size at most 0.9|U|0.9|U|.

A subtlety in addressing the final goal is that in the original analysis of [Li21], an argument is made that randomly sampling UUU^{\prime}\subset U will, with reasonable probability, mean that for some vv, its Min Isolating Cut SvS_{v} will be the same as its Min-ss-vv-Cut SvS^{*}_{v}. In particular, this will be true if vv is the only vertex sampled on its side of the Min-ss-vv-Cut. For the right sampling rate, this happens with probability O(1/|S|)O(1/|S^{*}|) but contributes |S||S^{*}| to the output size. So, in this case, each vertex in UU contributes constant expectation to the output size (up to log\log factors coming from choosing the right sampling rate and enforcing the second goal). Unfortunately, even though the Min Isolating Cuts output by our DP algorithm have small additive error, it is possible for them to be much smaller than the optimal |Sv||S^{*}_{v}| even if we sample the right terminals to get a set SvS_{v} with approximately the same cost as SvS^{*}_{v}. In order to make the analysis work with approximation, we give up on comparing to SvS^{*}_{v} and instead compare ourselves to the smallest cardinality set S~v\tilde{S}_{v} which is an approximate Min-ss-vv-Cut. For the full argument, we adjust our notion of approximation based on the size of S~v\tilde{S}_{v}, allowing for weaker approximations for smaller cardinality sets. This only degrades our approximation with respect to the first goal by an extra logarithmic factor.

The privacy guarantee.

The second challenge we face is controlling the privacy budget. We need to demonstrate that for two neighboring graphs, the distribution of the outputs of polylog(n)\operatorname{polylog}(n) recursive layers differs by at most a factor of eεe^{\varepsilon}. To achieve this, we split our privacy budget across polylog(n)\operatorname{polylog}(n) recursive sub-instances. We point out that the recursion depth of polylog(n)\operatorname{polylog}(n) does not directly imply that the privacy budget is split across polylog(n)\operatorname{polylog}(n) instances. We describe this challenge in more detail and the way our algorithm bypasses it.

To understand this more clearly, first consider two neighboring instances, GG and GG^{\prime}, that differ by the edge xyxy. Suppose that in the first step of the algorithm, the good sets SvS_{v} and SlargeS_{\text{large}} are the same in both graphs GG and GG^{\prime}. If both xx and yy are in one of the good sets SvS_{v} outputted by the first step, or if they are both in SlargeS_{\text{large}}, all respective recursion instances are exactly the same except for one. For example, if x,ySvx,y\in S_{v}, then in all instances where SvS_{v} is contracted, the resulting recursion graph does not depend on the edge xyxy and is, therefore, the same in both instances.

On the other hand, suppose that xSvx\in S_{v} and ySuy\in S_{u}. In this case, the edge xyxy influences multiple recursion instances: specifically, in GvG_{v}, GuG_{u}, and GlargeG_{\text{large}} (similarly GvG^{\prime}_{v}, GuG^{\prime}_{u}, and GlargeG^{\prime}_{\text{large}}). This poses a challenge because the computation in multiple branches of the same recursion call depends on the same edge. Consequently, the privacy guarantee is not solely affected by the recursion depth.

To overcome this, we add a simple step before recursion on some of the new instances. For instances obtained by contracting all vertices in VSvV\setminus S_{v} to a single vertex and recursing on the resulting graph, we first add an edge from the contracted vertex to every vertex in SvS_{v} with weights drawn from Lap(polylog(n)/ε)\text{Lap}(\operatorname{polylog}(n)/\varepsilon). Then we recurse on this altered graph. The intuition is that in neighboring instances, if xSvx\in S_{v} and ySvy\notin S_{v}, these noisy edges cancel the influence of xyxy, preventing its influence from propagating further down this branch of the recursion. Thus, xyxy only impacts GlargeG_{\text{large}}. Fortunately, the additional added noise only contributes O~(n/ε)\tilde{O}(n/\varepsilon) to the final additive error as noise is only added on O(n)O(n) edges.

1.3 Organization

We provide necessary definitions and basic lemmas in Section 2. We present our DP Min Isolating Cuts algorithm in Section 3 and prove Theorem 1.2. In Section 4 we present the recursive step we use in our DP GH-tree algorithm, and in Section 5 we present our final algorithm and prove Theorem 1.1.

1.4 Open Problems

Our algorithm outputs an O~(n/ε)\tilde{O}(n/\varepsilon)-approximate Gomory-Hu Tree from which we can recover an approximate Min-ss-tt-Cut for any two distinct vertices ss and tt. The additive error is necessary up to polylogn\operatorname{polylog}n and 1/ε1/\varepsilon factors by the lower bound in [DMN23] for just releasing a single Min-ss-tt-Cut.

It is an interesting open question whether one can get a better additive error or prove lower bounds in the setting where we are only required to release approximate values of the cuts and not the cuts themselves. To our knowledge, the best achievable error is O(n/ε)O(n/\varepsilon) under approximate DP via the trivial algorithm, which adds Lap(n/ε)\text{Lap}(n/\varepsilon) noise to each true value (this is private via “advanced composition” [DR14]). Prior to our work, the best algorithm in the pure DP setting was to solve the all-cuts/synthetic graph problem, incurring O(n3/2/ε)O(n^{3/2}/\varepsilon) error [GRU12]. Our work yields O~(n/ε)\tilde{O}(n/\varepsilon) error for this problem also with pure DP, but no non-trivial lower bound is known for releasing APMC values. The Ω(n)\Omega(n) lower bound of [DMN23] is for releasing an approximate Min-ss-tt-Cut, but releasing a single cut value can be done trivially with O(1/ε)O(1/\varepsilon) error via the Laplace mechanism. This question parallels the all-pairs shortest-paths distances problem studied in [Sea16, FLL22, CGK+23, BDG+24] for which sublinear additive error is possible.

Another open problem is whether there is a polynomial time ε\varepsilon-DP algorithm for the global Min Cut problem with error below O~(n/ε)\tilde{O}(n/\varepsilon). As mentioned below Corollary 1.2, the polynomial time algorithm from [GLM+10] is only approximate DP (but can be made pure DP if the runtime is exponential). Lastly, the same question can be asked for the minimum kk-cut problem (see Corollary 1.3): what are the limits of efficient, i.e., polynomial time, algorithms that are also pure DP?

2 Preliminaries

2.1 Notation

We will denote a weighted, undirected graph by G=(V,E,w)G=(V,E,w). For a subset of vertices SVS\subseteq V, we will use GS\partial_{G}S, or simply S\partial S, to denote the set of edges between SS and VSV\setminus S. For a set of edges QEQ\subseteq E, we will use w(Q)w(Q) to denote the sum of the weights of the edges in QQ. For a set of vertices SVS\subseteq V, we will use w(S)w(S) to denote w(S)w(\partial S), for simplicity. For s,tVs,t\in V, we use λG(s,t)\lambda_{G}(s,t) to denote the Min-ss-tt-Cut value in GG. We will assume that all Min-ss-tt-Cuts or Min Isolating Cuts are unique. For n>0n>0, lg(n)\lg(n) is logarithm base-22 and ln(n)\ln(n) is logarithm base-ee.

2.2 Graph Cuts

In our algorithms, we use the notion of vertex contractions which we formally define here.

Definition 2.1 (Vertex contractions).

Let XVX\subseteq V be a subset of vertices of the graph G=(V,E,w)G=(V,E,w). Contracting the set XX into a vertex is done as follows: we add a vertex xx to the graph and remove all the vertices in XX from the graph. Then for every vertex vVXv\in V\setminus X, we add an edge from xx to vv with weight xXw(xv)\sum_{x^{\prime}\in X}w(x^{\prime}v). Note that if none of the vertices in XX has an edge to vv, then there is no edge from xx to vv.

We use the submodularity property of cuts in many of our proofs.

Lemma 2.1 (Submodularity of Cuts [Cun85]).

For any graph G=(V,E,w)G=(V,E,w), and any two subsets S,TVS,T\subseteq V,

w(S)+w(T)w(ST)+w(ST).w(S)+w(T)\geq w(S\cup T)+w(S\cap T).

Recall the definition of Min Isolating Cuts problem (see Definition 1.1). We use the following simple fact.

Fact 2.1 ([LP20]).

There always exists a minimum isolating cuts solution where the solution {Sv:vR}\{S_{v}:v\in R\} are disjoint.

Definition 2.2 (Gomory-Hu Steiner tree [Li21]).

Given a graph G=(V,E,w)G=(V,E,w) and a set of terminals UVU\subseteq V, the Gomory-Hu Steiner tree is a weighted tree TT on the vertices UU, together with a function f:VUf:V\rightarrow U, such that For all s,tUs,t\in U, consider the minimum-weight edge uvuv on the unique ss-tt path in TT. Let U0U_{0} be the vertices of the connected component of TuvT-uv containing ss. Then, the set f1(U0)Vf^{-1}(U_{0})\subseteq V is a Min-ss-tt-Cut, and its value is wT(uv)w_{T}(uv).

Note that for U=VU=V and f(v)=vf(v)=v the Gomory-Hu Steiner tree equals the Gomory-Hu tree.

2.3 Differential Privacy

Definition 2.3 (Edge-Neighboring Graphs).

Graphs G=(V,E,w)G=(V,E,w) and G=(V,E,w)G^{\prime}=(V,E^{\prime},w^{\prime}) are called edge-neighboring if there is uvV2uv\in V^{2} such that |wG(uv)wG(uv)|1|w_{G}(uv)-w_{G^{\prime}}(uv)|\leq 1 and for all uvuvu^{\prime}v^{\prime}\neq uv, uvV2u^{\prime}v^{\prime}\in V^{2}, we have wG(uv)=wG(uv)w_{G}(u^{\prime}v^{\prime})=w_{G^{\prime}}(u^{\prime}v^{\prime}).

Definition 2.4 (Differential Privacy [Dwo06]).

A (randomized) algorithm 𝒜\mathcal{A} is (ε,δ)(\varepsilon,\delta)-private (or (ε,δ)(\varepsilon,\delta)-DP) if for any neighboring graphs GG and GG^{\prime} and any set of outcomes ORange(𝒜)O\subset Range(\mathcal{A}) it holds

[𝒜(G)O]eε[𝒜(G)O]+δ.\operatorname*{\mathbb{P}}\left[\mathcal{A}(G)\in O\right]\leq e^{\varepsilon}\operatorname*{\mathbb{P}}\left[\mathcal{A}(G^{\prime})\in O\right]+\delta.

When δ=0\delta=0, algorithm 𝒜\mathcal{A} is pure differentially private, or ε\varepsilon-DP.

Theorem 2.1 (Basic composition [DMNS06, DL09]).

Let ε1,,εt>0\varepsilon_{1},\ldots,\varepsilon_{t}>0 and δ1,,δt0\delta_{1},\ldots,\delta_{t}\geq 0. If we run tt (possibly adaptive) algorithms where the ii-th algorithm is (εi,δi)(\varepsilon_{i},\delta_{i})-DP, then the entire algorithm is (ε1++εt,δ1++δt)(\varepsilon_{1}+\ldots+\varepsilon_{t},\delta_{1}+\ldots+\delta_{t})-DP.

Theorem 2.2 (Laplace mechanism [DR14]).

Consider any function ff which maps graphs GG to d\mathbb{R}^{d} with the property that for any two neighboring graphs G,GG,G^{\prime}, |f(G)f(G)|Δ|f(G)-f(G^{\prime})|\leq\Delta. Then, releasing

f(G)+(X1,,Xd)f(G)+(X_{1},\ldots,X_{d})

where each XiX_{i} is i.i.d. with XiLap(Δ/ε)X_{i}\sim\text{Lap}(\Delta/\varepsilon) satisfies ε\varepsilon-DP.

Theorem 2.3 (Private Min-ss-tt-Cut [DMN23]).

Fix any ε>0\varepsilon>0. There is an (ε,0)(\varepsilon,0)-DP algorithm PrivateMin-s-t-Cut(G=(V,E,w),s,t)\text{PrivateMin-s-t-Cut}(G=(V,E,w),s,t) for stVs\neq t\in V that reports an ss-tt cut for nn-vertex weighted graphs that is within O(nε)O(\frac{n}{\varepsilon}) additive error from the Min-ss-tt-Cut with high probability.

By standard techniques, we can also use Theorem 2.3 to design an (ε,0)(\varepsilon,0)-DP algorithm for computing an approximate min-SS-TT-cut for two disjoint subsets S,TVS,T\subseteq V that is within O(nε)O(\frac{n}{\varepsilon}) additive error from the actual min-SS-TT-cut (e.g., by contracting all vertices in SS and all vertices in TT to two supernodes). Furthermore, our final algorithm is recursive with many calls to Private Min-SS-TT-Cut for graphs with few vertices, and it is not enough to succeed with high probability with respect to nn. The error analysis of [DMN23] shows that the error is bounded by the sum O(n)O(n) random variables distributed as Exp(ε)\text{Exp}(\varepsilon). Using Theorem 2.4 yields the following corollary:

Corollary 2.1 (Private Min-SS-TT-Cut).

Fix any ε>0\varepsilon>0, there exists an (ε,0)(\varepsilon,0)-DP algorithm
PrivateMin-S-T-Cut(G=(V,E,w),S,T,ε)\text{PrivateMin-S-T-Cut}(G=(V,E,w),S,T,\varepsilon) for disjoint S,TVS,T\subseteq V that reports a set CVC\subseteq V where SCS\subseteq C and CT=C\cap T=\emptyset, and w(C)w(\partial C) is within O\MT@delim@Auto\paren@starn+log(1/β)εO\MT@delim@Auto\paren@star{\frac{n+\log(1/\beta)}{\varepsilon}} additive error from the min-SS-TT-cut with probability at least 1β1-\beta.

2.4 Concentration Inequalities

Theorem 2.4 (Sums of Exponential Random Variables (Theorem 5.1 of [Jan17])).

Let X1,,XNX_{1},\ldots,X_{N} be independent random variables with XiExp(ai)X_{i}\sim\text{Exp}(a_{i}). Let μ=i=1N1ai\mu=\sum_{i=1}^{N}\frac{1}{a_{i}} be the expectation of the sum of the XiX_{i}’s and let a=miniaia^{*}=\min_{i}a_{i}. Then, for any λ1\lambda\geq 1,

[iSXiλμ]1λexp\MT@delim@Auto\brack@staraμ(λ1lnλ)\operatorname*{\mathbb{P}}\left[\sum_{i\in S}X_{i}\geq\lambda\mu\right]\leq\frac{1}{\lambda}\exp\MT@delim@Auto\brack@star{-a^{*}\mu(\lambda-1-\ln\lambda)}

3 Private Min Isolating Cuts

In this section we prove Theorem 1.2. In fact we prove a stronger version denoted in Lemma 3.1. The steps in the algorithm that differ meaningfully from the non-private version are in color.

Algorithm 1 PrivateMinIsolatingCuts(G=(V,E,w),R,U,ε,β)(G=(V,E,w),R,U,\varepsilon,\beta)
1:Initialize WrVW_{r}\leftarrow V for every rRr\in R
2:Identify RR with {0,,|R|1}\{0,\ldots,|R|-1\}
3:for ii from 0 to lg(|R|1)\lfloor\lg(|R|-1)\rfloor do
4:  Ai{rR:rmod2i+1<2i}A_{i}\leftarrow\{r\in R:r\bmod 2^{i+1}<2^{i}\}
5:  CiC_{i}\leftarrowPrivateMin-S-T-Cut(G,Ai,RAi,ε/(lg|R|+2))\text{PrivateMin-S-T-Cut}(G,A_{i},R\setminus A_{i},\varepsilon/(\lg|R|+2))
6:  WrWrCiW_{r}\leftarrow W_{r}\cap C_{i} for every rAir\in A_{i}
7:  WrWr(VCi)W_{r}\leftarrow W_{r}\cap(V\setminus C_{i}) for every rRAir\in R\setminus A_{i}
8:end for
9:for rRr\in R do
10:  Let HrH_{r} be GG with all vertices in VWrV\setminus W_{r} contracted, and let trt_{r} be the contracted vertex
11:  In HrH_{r}, add weight BH(n+lg(1/β))lg2(|R|)}ε|U|B_{\text{H}}{}\cdot\frac{(n+\lg(1/\beta))\lg^{2}(|R|)\}}{\varepsilon|U|} for every edge from vertex in WrUW_{r}\cap U to trt_{r} for some sufficiently large constant BHB_{\text{H}}{}.
12:end for
13:rRHr{\mathcal{H}}\leftarrow\bigcup_{r\in R}H_{r}
14:𝒞{\mathcal{C}}\leftarrowPrivateMin-S-T-Cut(,R,{tr}rR,ε/(lg|R|+2))\text{PrivateMin-S-T-Cut}({\mathcal{H}},R,\{t_{r}\}_{r\in R},\varepsilon/(\lg|R|+2))
15:return {𝒞Wr}rR\{{\mathcal{C}}\cap W_{r}\}_{r\in R}
Lemma 3.1.

On a graph GG with nn vertices, a set of terminals RVR\subseteq V, another set of vertices UVU\subseteq V, and a privacy parameter ε\varepsilon, there is an (ε,0)(\varepsilon,0)-DP algorithm PrivateMinIsolatingCuts(G,R,U,ε,β)(G,R,U,\varepsilon,\beta) that returns a set of Isolating Cuts over terminals RR. The total cut values of the Isolating Cuts is within additive error O((n+lg(1/β))lg2(|R|)/ε)O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon) from the Min Isolating Cuts with probability 1β1-\beta.

Furthermore, if the Min Isolating Cut for any terminal rRr\in R contains at most 0.5|U|0.5|U| vertices from UU, then the Isolating Cut for rr returned by the algorithm will contain at most 0.9|U|0.9|U| vertices from UU, with high probability.

Proof.

The algorithm is presented in Algorithm 1. On a high level, the algorithm follows the non-private Min Isolating Cuts algorithm by [LP20, AKT21], but replacing all calls to Min-SS-TT-Cut with private Min-SS-TT-Cut from Corollary 2.1. One added step is 11, which is used to provide the guarantee that if the Min Isolating Cut for terminal rr contains a small number of vertices in UU, then the isolating cut for terminal rr returned by the algorithm also does.

Next, we explain the algorithm in more detail. For every rRr\in R, we maintain a set WrW_{r} that should contain the rr-side of a (r,R{r})(r,R\setminus\{r\}) cut obtained in the algorithm. In each of the lg(|R|1)+1\lfloor\lg(|R|-1)\rfloor+1 iterations, we find a subset AiRA_{i}\subseteq R, and find a cut that separates AiA_{i} from RAiR\setminus A_{i}. Let CiC_{i} be the side of the cut containing AiA_{i}. Then for every rAir\in A_{i}, we update WrW_{r} with WrCiW_{r}\cap C_{i}; for every rRAir\in R\setminus A_{i}, we update WrW_{r} with Wr(VCi)W_{r}\cap(V\setminus C_{i}). The choice of AiA_{i} is so that every pair r1,r2Rr_{1},r_{2}\in R are on different sides of the Min-SS-TT Cut in at least one iteration; as a result, WrR={r}W_{r}\cap R=\{r\} for every rRr\in R after all iterations.

Next, for every rRr\in R, the algorithm aims to compute a cut separating rr from R{r}R\setminus\{r\}, where the side containing rr is inside WrW_{r}. This can be done by contracting all vertices outside of WrW_{r} to a vertex trt_{r} and computing private Min-rr-trt_{r} Cut. To incentivize cuts that contain fewer vertices in UU on the side containing rr, the algorithm adds a positive weight from every vertex in WrUW_{r}\cap U to trt_{r}. Finally, these private Min-rr-trt_{r} Cut instances can be solved at once by combining them into a single graph {\mathcal{H}}.

Privacy analysis.

The only parts of the algorithm that depend on the edges or edge weights are the calls to PrivateMin-S-T-Cut. Each call to PrivateMin-S-T-Cut is (ε/(lg|R|+2),0)(\varepsilon/(\lg|R|+2),0)-DP, and the number of calls is lg(|R|1)+2\lfloor\lg(|R|-1)\rfloor+2, so the overall algorithm is ε\varepsilon-DP via basic composition (Theorem 2.1).

Error analysis.

First, we analyze the error introduced by the for loop starting at 3. Let {Sr}rR\{S_{r}\}_{r\in R} be the (non-private) Min Isolating Cuts for terminals in RR, these are only used for analysis purposes. Take an iteration ii of the for loop and let {Wr}rR\{W_{r}\}_{r\in R} be the values of WrW_{r}’s before the start of the iteration, and let {Wr}rR\{W_{r}^{\prime}\}_{r\in R} denote the value of WrW_{r}’s at the end of the iteration. We show the following claim:

Claim 1.

With high probability, rRw(WrSr)rRw(WrSr)+O(nlg(|R|)/ε)\sum_{r\in R}w\left(W^{\prime}_{r}\cap S_{r}\right)\leq\sum_{r\in R}w\left(W_{r}\cap S_{r}\right)+O(n\lg(|R|)/\varepsilon).

Proof.

We first show rAiw(WrSr)rAiw(WrSr)+O(nlg(|R|)/ε)\sum_{r\in A_{i}}w\left(W^{\prime}_{r}\cap S_{r}\right)\leq\sum_{r\in A_{i}}w\left(W_{r}\cap S_{r}\right)+O(n\lg(|R|)/\varepsilon). Let SAi:=rAi(SrWr)S^{\prime}_{A_{i}}:=\bigcup_{r\in A_{i}}(S_{r}\cap W_{r}). By Lemma 2.1,

w(SAi)+w(Ci)w(SAiCi)+w(SAiCi).w(S^{\prime}_{A_{i}})+w(C_{i})\geq w(S^{\prime}_{A_{i}}\cup C_{i})+w(S^{\prime}_{A_{i}}\cap C_{i}). (1)

Recall that with probability 1β/lg|R|1-\beta/\lg|R|, CiC_{i} is within O((n+lg(lg|R|/β))lg(|R|)/ε)=O((n+lg(1/β))lg(|R|)/ε)O((n+\lg(\lg|R|/\beta))\lg(|R|)/\varepsilon)=O((n+\lg(1/\beta))\lg(|R|)/\varepsilon) of the minimum cut separating AiA_{i} and RAiR\setminus A_{i}, by the guarantee of Corollary 2.1, and note that SAiCiS^{\prime}_{A_{i}}\cup C_{i} is also a cut separating AiA_{i} and RAiR\setminus A_{i}. Therefore,

w(Ci)w(SAiCi)+O((n+lg(1/β))lg(|R|)/ε).w(C_{i})\leq w(S^{\prime}_{A_{i}}\cup C_{i})+O((n+\lg(1/\beta))\lg(|R|)/\varepsilon). (2)

Combining Equations 1 and 2, we get that

w(SAiCi)w(SAi)+O((n+lg(1/β))lg(|R|)/ε).w(S^{\prime}_{A_{i}}\cap C_{i})\leq w(S^{\prime}_{A_{i}})+O((n+\lg(1/\beta))\lg(|R|)/\varepsilon). (3)

Therefore,

rAiw(WrSr)\displaystyle\sum_{r\in A_{i}}w\left(W^{\prime}_{r}\cap S_{r}\right) =rAiw(WrSrCi)\displaystyle=\sum_{r\in A_{i}}w\left(W_{r}\cap S_{r}\cap C_{i}\right)
=w(rAi(WrSrCi))+r1r2Aiw(E((Wr1Sr1Ci)×(Wr2Sr2Ci)))\displaystyle=w\left(\bigcup_{r\in A_{i}}\left(W_{r}\cap S_{r}\cap C_{i}\right)\right)+\sum_{r_{1}\neq r_{2}\in A_{i}}w\left(E\cap\left(\left(W_{r_{1}}\cap S_{r_{1}}\cap C_{i}\right)\times\left(W_{r_{2}}\cap S_{r_{2}}\cap C_{i}\right)\right)\right)
w(SAiCi)+r1r2Aiw(E((Wr1Sr1)×(Wr2Sr2)))\displaystyle\leq w(S^{\prime}_{A_{i}}\cap C_{i})+\sum_{r_{1}\neq r_{2}\in A_{i}}w\left(E\cap\left(\left(W_{r_{1}}\cap S_{r_{1}}\right)\times\left(W_{r_{2}}\cap S_{r_{2}}\right)\right)\right)
w(SAi)+O((n+lg(1/β))lg(|R|)/ε)+r1r2Aiw(E((Wr1Sr1)×(Wr2Sr2)))\displaystyle\leq w(S^{\prime}_{A_{i}})+O((n+\lg(1/\beta))\lg(|R|)/\varepsilon)+\sum_{r_{1}\neq r_{2}\in A_{i}}w\left(E\cap\left(\left(W_{r_{1}}\cap S_{r_{1}}\right)\times\left(W_{r_{2}}\cap S_{r_{2}}\right)\right)\right) (by Equation 3)
=rAiw(WrSr)+O((n+lg(1/β))lg(|R|)/ε).\displaystyle=\sum_{r\in A_{i}}w\left(W_{r}\cap S_{r}\right)+O((n+\lg(1/\beta))\lg(|R|)/\varepsilon).

By an analogous argument, we can show rRAiw(WrSr)rRAiw(WrSr)+O(nlg(|R|)/ε)\sum_{r\in R\setminus A_{i}}w\left(W^{\prime}_{r}\cap S_{r}\right)\leq\sum_{r\in R\setminus A_{i}}w\left(W_{r}\cap S_{r}\right)+O(n\lg(|R|)/\varepsilon). Summing up the two inequalities gives the desired claim. \Box

By applying 1 repeatedly, we can easily show the following claim:

Claim 2.

At the end of the for loop starting at 3, rRw(WrSr)rRw(Sr)+O((n+lg(1/β))lg2(|R|)/ε)\sum_{r\in R}w(W_{r}\cap S_{r})\leq\sum_{r\in R}w(S_{r})+O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon), with probability 1β1-\beta.

The following claim is a simple observation:

Claim 3.

At the end of the for loop starting at 3, rWrr\in W_{r} for every rRr\in R and distinct WrW_{r}’s are disjoint.

Next, we show that the Min-rr-trt_{r} Cut values in HrH_{r} are close to the Min Isolating Cut values:

Claim 4.

With probability 1β1-\beta, rRλHr(r,tr)rRwG(Sr)+O((n+lg(1/β))lg2(|R|)/ε)\sum_{r\in R}\lambda_{H_{r}}(r,t_{r})\leq\sum_{r\in R}w_{G}(S_{r})+O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon).

Proof.

We have that

rRλHr(r,tr)\displaystyle\sum_{r\in R}\lambda_{H_{r}}(r,t_{r}) rRwHr(WrSr)\displaystyle\leq\sum_{r\in R}w_{H_{r}}(W_{r}\cap S_{r})
=rR(wG(WrSr)+|WrSrU|O((n+lg(1/β))lg2(|R|)/(ε|U|)))\displaystyle=\sum_{r\in R}\left(w_{G}(W_{r}\cap S_{r})+|W_{r}\cap S_{r}\cap U|\cdot O((n+\lg(1/\beta))\lg^{2}(|R|)/(\varepsilon|U|))\right)
rRwG(WrSr)+O((n+lg(1/β))lg2(|R|)/ε)\displaystyle\leq\sum_{r\in R}w_{G}(W_{r}\cap S_{r})+O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon)
rRwG(Sr)+O((n+lg(1/β))lg2(|R|)/ε),\displaystyle\leq\sum_{r\in R}w_{G}(S_{r})+O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon),

where the last step is by 2. \Box

Because of 4 (and note that |V()|=O(n)|V({\mathcal{H}})|=O(n)), with high probability, the final cuts 𝒞Wr{\mathcal{C}}\cap W_{r} returned by the algorithm will have the property that

rRwHr(𝒞Wr)\displaystyle\sum_{r\in R}w_{H_{r}}({\mathcal{C}}\cap W_{r}) rRλHr(r,tr)+O((n+lg(1/β))lg(|R|)/ε)\displaystyle\leq\sum_{r\in R}\lambda_{H_{r}}(r,t_{r})+O((n+\lg(1/\beta))\lg(|R|)/\varepsilon)
rRwG(Sr)+O((n+lg(1/β))lg2(|R|)/ε).\displaystyle\leq\sum_{r\in R}w_{G}(S_{r})+O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon).

Furthermore, wG(𝒜Wr)wHr(𝒜Wr)w_{G}({\mathcal{A}}\cap W_{r})\leq w_{H_{r}}({\mathcal{A}}\cap W_{r}) as we only add positive weights to HrH_{r} compared to GG, we further get

rRwG(𝒞Wr)rRwG(Sr)+O((n+lg(1/β))lg2(|R|)/ε),\sum_{r\in R}w_{G}({\mathcal{C}}\cap W_{r})\leq\sum_{r\in R}w_{G}(S_{r})+O((n+\lg(1/\beta))\lg^{2}(|R|)/\varepsilon),

which is the desired error bound.

Additional guarantee.

Finally, we need to show that if |SrU|0.5U|S_{r}\cap U|\leq 0.5U for some rRr\in R, then with high probability, the returned isolating cut by the algorithm 𝒞Wr{\mathcal{C}}\cap W_{r} has |𝒞WrU|0.9U|{\mathcal{C}}\cap W_{r}\cap U|\leq 0.9U. By Corollary 2.1,

w(𝒞)w((𝒞Wr)(SrWr))+O((n+lg(1/β))(lg(|R|))/ε),w_{{\mathcal{H}}}({\mathcal{C}})\leq w_{{\mathcal{H}}}(({\mathcal{C}}\setminus W_{r})\cup(S_{r}\cap W_{r}))+O((n+\lg(1/\beta))(\lg(|R|))/\varepsilon),

with high probability. By removing cut values contributed by HrH_{r^{\prime}} for rrr^{\prime}\neq r from both sides, we get that

wHr(𝒞Wr)wHr(SrWr)+O((n+lg(1/β))(lg(|R|))/ε).w_{H_{r}}({\mathcal{C}}\cap W_{r})\leq w_{H_{r}}(S_{r}\cap W_{r})+O((n+\lg(1/\beta))(\lg(|R|))/\varepsilon).

Rewriting the cut values in terms of the edge weights of GG instead of HrH_{r}, the above becomes

w(𝒞Wr)+(BH(n+lg(1/β))lg2(|R|)/(ε|U|))|𝒞WrU|\displaystyle w({\mathcal{C}}\cap W_{r})+(B_{\text{H}}{}\cdot(n+\lg(1/\beta))\lg^{2}(|R|)/(\varepsilon|U|))|{\mathcal{C}}\cap W_{r}\cap U|
w(SrWr)+(BH(n+lg(1/β))lg2(|R|)/(ε|U|))|SrWrU|+O((n+lg(1/β))(lg(|R|))/ε)\displaystyle\leq w(S_{r}\cap W_{r})+(B_{\text{H}}{}\cdot(n+\lg(1/\beta))\lg^{2}(|R|)/(\varepsilon|U|))|S_{r}\cap W_{r}\cap U|+O((n+\lg(1/\beta))(\lg(|R|))/\varepsilon)
w(Sr)+(BH(n+lg(1/β))lg2(|R|)/(ε|U|))|SrWrU|+O((n+lg(1/β))(lg2(|R|))/ε).\displaystyle\leq w(S_{r})+(B_{\text{H}}{}\cdot(n+\lg(1/\beta))\lg^{2}(|R|)/(\varepsilon|U|))|S_{r}\cap W_{r}\cap U|+O((n+\lg(1/\beta))(\lg^{2}(|R|))/\varepsilon). (By 2)

Because w(𝒞Wr)w(Sr)w({\mathcal{C}}\cap W_{r})\geq w(S_{r}), the above implies

|𝒞WrU|\displaystyle|{\mathcal{C}}\cap W_{r}\cap U| O((n+lg(1/β))(lg2(|R|))/ε)BH(n+lg(1/β))lg2(|R|)/(ε|U|)+|SrWrU|\displaystyle\leq\frac{O((n+\lg(1/\beta))(\lg^{2}(|R|))/\varepsilon)}{B_{\text{H}}{}\cdot(n+\lg(1/\beta))\lg^{2}(|R|)/(\varepsilon|U|)}+|S_{r}\cap W_{r}\cap U|
0.4|U|+|SrWrU|\displaystyle\leq 0.4|U|+|S_{r}\cap W_{r}\cap U| (By setting BHB_{\text{H}}{} large enough)
0.4|U|+|SrU|\displaystyle\leq 0.4|U|+|S_{r}\cap U|
0.9|U|,\displaystyle\leq 0.9|U|,

as desired. \Box

4 Core Recursive Step

Algorithm 2 PrivateGHTreeStep(G=(V,E,w),s,U,ε,β)(G=(V,E,w),s,U,{\color[rgb]{0.8,0.25,0.33}{\varepsilon,\beta}})
1:ΓisoO\MT@delim@Auto\paren@star(n+lg(1/β))lg3(|U|)ε{\Gamma_{\text{iso}}}\leftarrow O\MT@delim@Auto\paren@star{\frac{(n+\lg(1/\beta))\lg^{3}(|U|)}{\varepsilon}} and ΓvaluesO\MT@delim@Auto\paren@star|U|lg(|U|/β)ε{\Gamma_{\text{values}}}\leftarrow O\MT@delim@Auto\paren@star{\frac{|U|\lg(|U|/\beta)}{\varepsilon}}
2:λ^(s,v)λ(s,v)+Lap\MT@delim@Auto\paren@star4(|U|1)ε\hat{\lambda}(s,v)\leftarrow\lambda(s,v)+\text{Lap}\MT@delim@Auto\paren@star{\frac{4(|U|-1)}{\varepsilon}} for all vU{s}v\in U\setminus\{s\}
3:Initialize R0UR^{0}\leftarrow U and DD\leftarrow\emptyset
4:for ii from 0 to \MT@delim@Auto\floor@starlg|U|\MT@delim@Auto\floor@star{\lg|U|} do
5:  Call PrivateMinIsolatingCuts\MT@delim@Auto\paren@starG,Ri,ε2(\MT@delim@Auto\floor@starlgU+1),β\MT@delim@Auto\floor@starlgU+1\MT@delim@Auto\paren@star{G,R^{i},\frac{\varepsilon}{2(\MT@delim@Auto\floor@star{\lg U}+1)},\frac{\beta}{\MT@delim@Auto\floor@star{\lg U}+1}} (Algorithm 1) obtaining disjoint sets S^vi{\hat{S}}_{v}^{i}
6:  w^(S^vi)w(S^vi)+Lap\MT@delim@Auto\paren@star8(\MT@delim@Auto\floor@starlgU+1)ε{\hat{w}}({\hat{S}}_{v}^{i})\leftarrow w({\hat{S}}_{v}^{i})+\text{Lap}\MT@delim@Auto\paren@star{\frac{8(\MT@delim@Auto\floor@star{\lg U}+1)}{\varepsilon}} for each vRi{s}v\in R^{i}\setminus\{s\}
7:  Let DiUD^{i}\subseteq U be the union of S^viU{\hat{S}}_{v}^{i}\cap U over all vRi{s}v\in R^{i}\setminus\{s\} satisfying w^(S^vi)λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg|U|i)+1Γiso+Γvalues{\hat{w}}({\hat{S}}_{v}^{i})\leq\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg|U|}-i)+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}} and |S^viU|(9/10)|U||{\hat{S}}_{v}^{i}\cap U|\leq(9/10)|U|
8:  Ri+1R^{i+1}\leftarrow sample of UU where each vertex in U{s}U\setminus\{s\} is sampled independently with probability 2i+12^{-i+1}, and ss is sampled with probability 11
9:end for
10:return the largest set DiD^{i}, the corresponding terminals vRi{s}v\in R^{i}\setminus\{s\} and sets S^vi{\hat{S}}_{v}^{i} satisfying the conditions on 7

We now describe a key subroutine, outlined as Algorithm 2, used to compute a DP Gomory-Hu tree. The high-level goal is to use Min Isolating Cuts to find minimum cuts that cover a large fraction of vertices in the graph. The overall structure of this algorithm follows that of the prior work [Li21] with several key changes to handle additive approximations and privacy. The inputs to Algorithm 2 are the weighted graph, a source vertex ss, a set of active vertices UVU\subseteq V, a privacy parameter ε\varepsilon, and a failure probability β\beta. The steps that differ meaningfully from the non-private version developed in [Li21] are in color. To obtain a DP version of this method, Algorithm 2 invokes DP Min-ss-tt-Cut and DP Min Isolating Cuts algorithm; the latter primitive is developed in this work in Section 3. In the original non-private algorithm, isolating cuts SviS_{v}^{i} are included in DiD^{i} if the set SviS_{v}^{i} corresponds to the vv side of a Min-ss-vv-Cut, i.e., w(Svi)=λ(s,v)w(S^{i}_{v})=\lambda(s,v). The analysis in prior work relies on this equality, i.e., on w(Svi)w(S^{i}_{v}) and λ(s,v)\lambda(s,v) being the same, in a crucial way. Informally speaking, it enables the selection of many Min Isolating Cuts of the right size. In our case, since the cuts and their values are released privately by random perturbations, it is unclear how to test that condition with equality. On the other hand, we still would like to ensure that many isolating cuts have “the right” size. Among our key technical contributions is relaxing that condition by using a condition which changes from iteration to iteration of the for-loop. The actual condition we use is

w^(S^vi)λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlgUi)+1Γiso+Γvalues{\hat{w}}({\hat{S}}_{v}^{i})\leq\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg U}-i)+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}} (4)

on 7 of Algorithm 2 where Γiso{\Gamma_{\text{iso}}} and Γvalues{\Gamma_{\text{values}}} are upper bounds on the additive errors of the approximate Min Isolating Cuts and the approximate Min-ss-vv-Cut values, respectively.

When using Equation 4, we also have to ensure that significant progress can still be made, i.e., to ensure that both (a) we will find a large set DiD^{i} which is the union of approximate Min Isolating Cuts S^vi{\hat{S}}_{v}^{i} satisfying the condition above and (b) none of the individual S^{\hat{S}} which we return are too large as we will recurse within each of these sets. A new analysis uses this changing inequality to show that the former is true. For the latter, we utilize the special property of our PrivateMinIsolatingCuts in Section 3 which forces an approximate isolating cut to contain at most 0.9|U|0.9|U| terminals if there exists an exact isolating cut of size at most |U|/2|U|/2. We now turn to the analysis.

4.1 Correctness

As in prior work [Li21, AKL+22], let DU{s}D^{*}\subseteq U\setminus\{s\} be the set of vertices vv such that if SvS_{v}^{*} is the vv side of the Min-ss-vv-Cut, |SvU||U|/2|S_{v}^{*}\cap U|\leq|U|/2.

Lemma 4.1.

PrivateGHTreeStep(G,U,s,ε,β)(G,U,s,\varepsilon,\beta) (Algorithm 2) has the following properties:

  • Let Γiso=C1(n+lg(1/β))lg3(|U|)/ε{\Gamma_{\text{iso}}}=C_{1}(n+\lg(1/\beta))\lg^{3}(|U|)/\varepsilon and Γvalues=C2|U|lg(|U|/β)/ε{\Gamma_{\text{values}}}=C_{2}|U|\lg(|U|/\beta)/\varepsilon for large enough constants C1,C2C_{1},C_{2}. Let {Svi}vRi\{S_{v}^{i}\}_{v\in R^{i}} be optimal Min Isolating Cuts for terminals RiR^{i}. Then, with probability at least 1O(β)1-O(\beta), the sets {S^vi:vR}\{{\hat{S}}_{v}^{i}:v\in R^{*}\} returned by the algorithm are approximate Min Isolating Cuts and approximate Min-vv-ss-Cuts:

    vRw(S^vi)w(Svi)Γiso,\sum_{v\in R^{*}}w({\hat{S}}_{v}^{i})-w(S_{v}^{i})\leq{\Gamma_{\text{iso}}},

    and, for all vRv\in R^{*},

    w(S^vi)λ(s,v)2(\MT@delim@Auto\floor@starlgU+1)Γiso+2Γvalues.w({\hat{S}}_{v}^{i})-\lambda(s,v)\leq 2(\MT@delim@Auto\floor@star{\lg U}+1){\Gamma_{\text{iso}}}+2{\Gamma_{\text{values}}}.
  • With probability at least 1/21/2, DiD^{i} returned by the algorithm satisfies

    |Di|=Ω(|D|lg|U|).|D^{i}|=\Omega\left(\frac{|D^{*}|}{\lg|U|}\right).

To prove this, we will need the following helpful definition and lemma. Let XviX_{v}^{i} be a random variable for the number of vertices in UU added to DiD^{i} by a set S^vi{\hat{S}}_{v}^{i}:

Xvi={|S^viU| if vRi and |S^viU|(9/10)|U| and w^(S^vi)λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg|U|i)+1Γiso+Γvalues0o.w. .X_{v}^{i}=\begin{cases}|{\hat{S}}_{v}^{i}\cap U|\qquad&\text{ if }v\in R^{i}\text{ and }|{\hat{S}}_{v}^{i}\cap U|\leq(9/10)|U|\text{ and }\\ &\quad{\hat{w}}({\hat{S}}_{v}^{i})\leq\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg|U|}-i)+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}\\ 0\qquad&\text{o.w. }\end{cases}. (5)
Lemma 4.2.

Consider a vertex vDv\in D^{*}. Let SviS_{v}^{i} be the vv part of an optimal solution to Min Isolating Cuts at stage ii. Assume that |λ(s,v)λ^(s,v)|Γvalues|\lambda(s,v)-\hat{\lambda}(s,v)|\leq{\Gamma_{\text{values}}} and |w(S^vi)w^(S^vi)|+|w(Svi)w(S^vi)|Γiso|w({\hat{S}}_{v}^{i})-{\hat{w}}({\hat{S}}_{v}^{i})|+|w(S_{v}^{i})-w({\hat{S}}_{v}^{i})|\leq{\Gamma_{\text{iso}}} for all i{0,\MT@delim@Auto\floor@starlg|U|}i\in\{0,\ldots\MT@delim@Auto\floor@star{\lg|U|}\}. Then, there exists an i{0,\MT@delim@Auto\floor@starlg|U|}i^{\prime}\in\{0,\ldots\MT@delim@Auto\floor@star{\lg|U|}\} such that

𝔼[Xvi]=O(1).{\mathbb{E}}\left[X_{v}^{i^{\prime}}\right]=O(1).
Proof.

Consider a specific sampling level i{0,,lg|U|}i\in\{0,\ldots,\lfloor\lg|U|\rfloor\}. We say that ii is “active” if there exists a set S~viU\tilde{S}_{v}^{i}\subset U containing vv and not ss such that |S~viU|[2i,2i+1)|\tilde{S}_{v}^{i}\cap U|\in[2^{i},2^{i+1}) and

w(S~vi)λ^(s,v)+2(\MT@delim@Auto\floor@starlg(|U|)i)Γiso+Γvalues.w(\tilde{S}_{v}^{i})\leq\hat{\lambda}(s,v)+2(\MT@delim@Auto\floor@star{\lg(|U|)}-i){\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}. (6)

Note that this is a deterministic property regarding the existence of such a set S~vi\tilde{S}_{v}^{i} independent of the randomness used to sample terminals or find private Min Isolating Cuts.

Let ii^{\prime} be the smallest active ii. Let SvS^{*}_{v} be the vv side of the true min s-vs\text{-}v cut, and let i=lg|SvU|i^{*}=\lfloor\lg|S^{*}_{v}\cap U|\rfloor. As w(Sv)=λ(s,v)λ^(s,v)+Γvaluesw(S^{*}_{v})=\lambda(s,v)\leq\hat{\lambda}(s,v)+{\Gamma_{\text{values}}}, ii^{*} must be active, so ii^{\prime} is well-defined and iii^{\prime}\leq i^{*}. As ii^{\prime} is active, there exists a set S~vi\tilde{S}_{v}^{i^{\prime}} with |S~viU|[2i,2i+1)|\tilde{S}_{v}^{i^{\prime}}\cap U|\in[2^{i^{\prime}},2^{i^{\prime}+1}) and with cost within 2(\MT@delim@Auto\floor@starlg(|U|)i)Γiso+Γvalues2(\MT@delim@Auto\floor@star{\lg(|U|)}-i^{\prime}){\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}} of λ^(s,v)\hat{\lambda}(s,v). On the other hand, as ii^{\prime} is the smallest active level, there is no set of size less than 2i2^{i^{\prime}} with cost within 2(\MT@delim@Auto\floor@starlg(|U|)(i1))Γiso+Γvalues2(\MT@delim@Auto\floor@star{\lg(|U|)}-(i^{\prime}-1)){\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}} of λ^(s,v)\hat{\lambda}(s,v).

Consider the event that in RiR^{i^{\prime}}, we sample vv but no other vertices in S~vi\tilde{S}_{v}^{i^{\prime}} as terminals, i.e., RiS~vi={v}R^{i^{\prime}}\cap\tilde{S}_{v}^{i^{\prime}}=\{v\}. Then, S~vi\tilde{S}_{v}^{i^{\prime}} would be a valid output of a call to isolating cuts. By the assumed guarantee of the error of the private Min Isolating Cuts algorithm, the actual cut we output has approximated cost:

w^(S^vi)\displaystyle{\hat{w}}({\hat{S}}_{v}^{i^{\prime}}) w(S~vi)+|w(S~vi)w^(S^vi)|\displaystyle\leq w(\tilde{S}_{v}^{i^{\prime}})+|w(\tilde{S}_{v}^{i^{\prime}})-{\hat{w}}({\hat{S}}_{v}^{i^{\prime}})|
w(S~vi)+|w(S~vi)w(S^vi)|+|w(S^vi)w^(S^vi)|\displaystyle\leq w(\tilde{S}_{v}^{i^{\prime}})+|w(\tilde{S}_{v}^{i^{\prime}})-w({\hat{S}}_{v}^{i^{\prime}})|+|w({\hat{S}}_{v}^{i^{\prime}})-{\hat{w}}({\hat{S}}_{v}^{i^{\prime}})|
w(S~vi)+Γiso\displaystyle\leq w(\tilde{S}_{v}^{i^{\prime}})+{\Gamma_{\text{iso}}}
λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg(|U|)i)+1Γiso+Γvalues.\displaystyle\leq\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg(|U|)}-i^{\prime})+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}.

For sake of contradiction, consider the case that |S^viU|<2i|{\hat{S}}_{v}^{i^{\prime}}\cap U|<2^{i^{\prime}}. Using the fact that all solutions of this size have large cost, we can conclude that

w^(S^vi)\displaystyle{\hat{w}}({\hat{S}}_{v}^{i^{\prime}}) w(S^vi)Γiso\displaystyle\geq w({\hat{S}}_{v}^{i^{\prime}})-{\Gamma_{\text{iso}}}
>λ^(s,v)+2(\MT@delim@Auto\floor@starlg(|U|)(i1))Γiso+ΓvaluesΓiso\displaystyle>\hat{\lambda}(s,v)+2(\MT@delim@Auto\floor@star{\lg(|U|)}-(i^{\prime}-1)){\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}-{\Gamma_{\text{iso}}}
>λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg(|U|)i))+1Γiso+Γvalues.\displaystyle>\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg(|U|)}-i^{\prime}))+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}.

This contradicts the previous inequality that shows that w^(S^vi){\hat{w}}({\hat{S}}_{v}^{i^{\prime}}) is upper bounded by this quantity, so |S^viU|2i|{\hat{S}}_{v}^{i^{\prime}}\cap U|\geq 2^{i^{\prime}} as long as the sampling event occurs.

Next, we show that |S^viU|(9/10)|U||{\hat{S}}_{v}^{i^{\prime}}\cap U|\leq(9/10)|U|. As vDv\in D^{*}, the true minimum cut SvS^{*}_{v} has the property |SvU||U|/2|S^{*}_{v}\cap U|\leq|U|/2. Furthermore, by the isolating cuts lemma of [LP20], the minimum isolating cut solution for any set of terminals including vv and ss will have that the vv part SvS_{v} is a subset of SvS^{*}_{v} (this is the basis for the isolating cuts algorithm). So, if vv is sampled in RiR^{i}, there will exist an optimal isolating cuts solution SviS_{v}^{i} with |SviU||U|/2|S_{v}^{i}\cap U|\leq|U|/2. By the guarantee of Lemma 3.1, |S^viU|(9/10)|U||{\hat{S}}_{v}^{i^{\prime}}\cap U|\leq(9/10)|U|.

Overall, we can bound the contribution of S^vi{\hat{S}}_{v}^{i^{\prime}} to DiD^{i^{\prime}} as

𝔼[Xvi]\displaystyle{\mathbb{E}}\left[X_{v}^{i^{\prime}}\right] |SviU|[v is the only vertex sampled in S~vi under sampling probability 2i]\displaystyle\geq|S_{v}^{i^{\prime}}\cap U|\cdot\operatorname*{\mathbb{P}}\left[\text{$v$ is the only vertex sampled in }\tilde{S}_{v}^{i^{\prime}}\text{ under sampling probability }2^{-i^{\prime}}\right]
2i(2i)(12i)|S~vi|1\displaystyle\geq 2^{i^{\prime}}\left(2^{-i^{\prime}}\right)\left(1-2^{-i^{\prime}}\right)^{|\tilde{S}_{v}^{i^{\prime}}|-1}
(12i)2i+12.\displaystyle\geq\left(1-2^{-i^{\prime}}\right)^{2^{i^{\prime}+1}-2}.

If i=0i^{\prime}=0, this evaluates to 11. Otherwise, if i1i^{\prime}\geq 1,

𝔼[Xvi](12i)2i+1=\MT@delim@Auto\paren@star11+2i12i2i+1\MT@delim@Auto\paren@star1e2i12i2i+1=e212ie4.{\mathbb{E}}\left[X_{v}^{i^{\prime}}\right]\geq\left(1-2^{-i^{\prime}}\right)^{2^{i^{\prime}+1}}=\MT@delim@Auto\paren@star{\frac{1}{1+\frac{2^{-i^{\prime}}}{1-2^{-i^{\prime}}}}}^{2^{i^{\prime}+1}}\geq\MT@delim@Auto\paren@star{\frac{1}{e^{\frac{2^{-i^{\prime}}}{1-2^{-i^{\prime}}}}}}^{2^{i^{\prime}+1}}=e^{-\frac{2}{1-2^{-i^{\prime}}}}\geq e^{-4}.

\Box

We are now ready to prove the main lemma of this section.

Proof of Lemma 4.1.

The first step of the proof will be to show that Γiso{\Gamma_{\text{iso}}} and Γvalues{\Gamma_{\text{values}}} upper bound the error of the approximate isolating cuts and min cut values used in the algorithm with probability 1O(β)1-O(\beta). Applying the guarantee of Lemma 3.1 and union bounding over all ii, with probability 1β1-\beta, we get the following guarantee of the quality of S^vi{\hat{S}}_{v}^{i}. If {Svi}\{S_{v}^{i}\} are optimal Min Isolating Cuts for terminals RiR^{i}:

vRiw(S^vi)w(Svi)O\MT@delim@Auto\paren@star(n+lg(1/β))lg3(|Ri|)ε.\sum_{v\in R^{i}}w({\hat{S}}_{v}^{i})-w(S_{v}^{i})\leq O\MT@delim@Auto\paren@star{\frac{(n+\lg(1/\beta))\lg^{3}(|R^{i}|)}{\varepsilon}}.

Furthermore, via the tail of the Laplace distribution and a union bound over all ii, each of the approximated weights satisfies the following inequality with probability 1β1-\beta:

|w^(S^vi)w(S^vi)|O\MT@delim@Auto\paren@starlg|U|(lg(|U|/β))ε.|{\hat{w}}({\hat{S}}_{v}^{i})-w({\hat{S}}_{v}^{i})|\leq O\MT@delim@Auto\paren@star{\frac{\lg|U|(\lg(|U|/\beta))}{\varepsilon}}.

Therefore, there exists a choice of Γiso=O\MT@delim@Auto\paren@star(n+lg(1/β))lg3(n)ε{\Gamma_{\text{iso}}}=O\MT@delim@Auto\paren@star{\frac{(n+\lg(1/\beta))\lg^{3}(n)}{\varepsilon}} such that with probability 1O(β)1-O(\beta), for all i{0,,\MT@delim@Auto\floor@starlg|U|}i\in\{0,\ldots,\MT@delim@Auto\floor@star{\lg|U|}\},

vRi|w(Svi)w(S^vi)|+|w(S^vi)w^(S^vi)|Γiso.\sum_{v\in R^{i}}|w(S_{v}^{i})-w({\hat{S}}_{v}^{i})|+|w({\hat{S}}_{v}^{i})-{\hat{w}}({\hat{S}}_{v}^{i})|\leq{\Gamma_{\text{iso}}}.

This satisfies the approximate Min Isolating Cuts guarantee of the lemma.

For the approximate minimum cut values, by the tail of the Laplace distribution and a union bound, each λ^(s,v)\hat{\lambda}(s,v) satisfies

|λ(s,v)λ^(s,v)|Γvalues=O\MT@delim@Auto\paren@star|U|lg(|U|/β)ε|\lambda(s,v)-\hat{\lambda}(s,v)|\leq{\Gamma_{\text{values}}}=O\MT@delim@Auto\paren@star{\frac{|U|\lg(|U|/\beta)}{\varepsilon}}

with probability 1β1-\beta. We will condition on these events going forward.

Sets S^vi{\hat{S}}_{v}^{i} are only included in our output if they are close to the min cut value λ(s,v)\lambda(s,v). Specifically, for any set returned by our algorithm:

w^(S^vi)λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg|U|i)+1Γiso+Γvalues.{\hat{w}}({\hat{S}}_{v}^{i})\leq\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg|U|}-i)+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}.

Applying the error guarantees for w^(S^vi){\hat{w}}({\hat{S}}_{v}^{i}), S^vi{\hat{S}}_{v}^{i}, and λ^(s,v)\hat{\lambda}(s,v),

w(S^vi)\displaystyle w({\hat{S}}_{v}^{i}) Γiso+w^(S^vi)\displaystyle\leq{\Gamma_{\text{iso}}}+{\hat{w}}({\hat{S}}_{v}^{i})
Γiso+λ^(s,v)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg|U|i)+1Γiso+Γvalues\displaystyle\leq{\Gamma_{\text{iso}}}+\hat{\lambda}(s,v)+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg|U|}-i)+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}
Γiso+(λ(s,v)+Γvalues)+\MT@delim@Auto\paren@star2(\MT@delim@Auto\floor@starlg|U|i)+1Γiso+Γvalues\displaystyle\leq{\Gamma_{\text{iso}}}+(\lambda(s,v)+{\Gamma_{\text{values}}})+\MT@delim@Auto\paren@star{2(\MT@delim@Auto\floor@star{\lg|U|}-i)+1}{\Gamma_{\text{iso}}}+{\Gamma_{\text{values}}}
λ(s,v)+2(\MT@delim@Auto\floor@starlg|U|+1)Γiso+2Γvalues.\displaystyle\leq\lambda(s,v)+2(\MT@delim@Auto\floor@star{\lg|U|}+1){\Gamma_{\text{iso}}}+2{\Gamma_{\text{values}}}.

This completes the first part of the proof concerning the error of the returned sets. In the remainder, we focus on the cardinality of the output.

By definition of XviX_{v}^{i}, the size of DiD^{i} is given by a sum over XviX_{v}^{i}:

|Di|=vU{s}Xvi|D^{i}|=\sum_{v\in U\setminus\{s\}}X_{v}^{i}

By linearity of expectation and as DU{s}D^{*}\subseteq U\setminus\{s\},

𝔼[i=0lg|U||Di|]i=0lg|U|vD𝔼[Xvi].{\mathbb{E}}\left[\sum_{i=0}^{\lfloor\lg|U|\rfloor}|D^{i}|\right]\geq\sum_{i=0}^{\lfloor\lg|U|\rfloor}\sum_{v\in D^{*}}{\mathbb{E}}\left[X_{v}^{i}\right].

As we output the largest DiD^{i} across all ii, the output of our algorithm will have expected size at least

1lg|U|+1i=0lg|U|vD𝔼[Xvi].\frac{1}{\lfloor\lg|U|\rfloor+1}\sum_{i=0}^{\lfloor\lg|U|\rfloor}\sum_{v\in D^{*}}{\mathbb{E}}\left[X_{v}^{i}\right].

Via Lemma 4.2 (note that the error condition holds with probability 1O(β)1-O(\beta) from the first part of this proof), the expected output size will be at least

Ω\MT@delim@Auto\paren@star|D|lg|U|.\Omega\MT@delim@Auto\paren@star{\frac{|D^{*}|}{\lg|U|}}.

The final result follows from Markov’s inequality. \Box

4.2 Privacy

Lemma 4.3.

PrivateGHTreeStep (Algorithm 2) is ε\varepsilon-DP.

Proof.

The algorithm PrivateGHTreeStep interacts with the sensitive edges only through calculations of approximate min cut values λ^(s,v)\hat{\lambda}(s,v), approximate isolating cut values w^(S^vi){\hat{w}}({\hat{S}}_{v}^{i}), and calls to PrivateMinIsolatingCuts (Algorithm 1). Otherwise, all computation only deals with the vertices of the graph which are public. The calculation of each of the |U|1|U|-1 cut values is ε4(|U|1)\frac{\varepsilon}{4(|U|-1)}-DP via the Laplace mechanism (Theorem 2.2). Via the privacy of PrivateMinIsolatingCuts, each call to that subroutine is ε2(\MT@delim@Auto\floor@starlg|U|+1)\frac{\varepsilon}{2(\MT@delim@Auto\floor@star{\lg|U|}+1)}-DP. At any sampling level ii, the approximate isolating cuts S^vi{\hat{S}}_{v}^{i} are disjoint. So, calculation of w(S^vi)w({\hat{S}}_{v}^{i}) for all vRi{s}v\in R^{i}\setminus\{s\} has sensitivity 22 as each edge can cross at most two sets. By the Laplace mechanism, calculation of all w^(S^vi){\hat{w}}({\hat{S}}_{v}^{i}) for any ii is ε4(\MT@delim@Auto\floor@starlg|U|+1)\frac{\varepsilon}{4(\MT@delim@Auto\floor@star{\lg|U|}+1)}-DP. Summing over all sampling levels, the total privacy via basic composition (Theorem 2.1) is:

(|U|1)ε4(|U|1)+(\MT@delim@Auto\floor@starlg|U|+1)\MT@delim@Auto\paren@starε2(\MT@delim@Auto\floor@starlg|U|+1)+ε4(\MT@delim@Auto\floor@starlg|U|+1)=ε4+ε2+ε4=ε.(|U|-1)\frac{\varepsilon}{4(|U|-1)}+(\MT@delim@Auto\floor@star{\lg|U|}+1)\MT@delim@Auto\paren@star{\frac{\varepsilon}{{2(\MT@delim@Auto\floor@star{\lg|U|}+1)}}+\frac{\varepsilon}{4(\MT@delim@Auto\floor@star{\lg|U|}+1)}}=\frac{\varepsilon}{4}+\frac{\varepsilon}{2}+\frac{\varepsilon}{4}=\varepsilon.

\Box

5 Final Algorithm

Algorithm 3 PrivateGHTree(G=(V,E,w),ε)(G=(V,E,w),\varepsilon)
1:(T,f)PrivateGHSteinerTree(G,V,ε/2,0,n)(T,f)\leftarrow{\color[rgb]{0.8,0.25,0.33}{\text{PrivateGHSteinerTree}{}(G,V,\varepsilon/2,0,n)}}
2:Add Lap\MT@delim@Auto\paren@star2(n1)ε\text{Lap}\MT@delim@Auto\paren@star{\frac{2(n-1)}{\varepsilon}} noise to each edge in TT
3:return TT
Algorithm 4 PrivateGHSteinerTree(G=(V,E,w),U,ε,t,nmax)(G=(V,E,w),U,{\color[rgb]{0.8,0.25,0.33}{\varepsilon,t,n_{\max}}})
1:tmaxΩ(lg2nmax)t_{\max}\leftarrow\Omega(\lg^{2}n_{\max})
2:if t>tmaxt>t_{\max} then
3:  return abort // the privacy budget is exhausted
4:end if
5:ss\leftarrow uniformly random vertex in UU
6:Call PrivateGHTreeStep(G,s,U,ε4tmax,1nmax3)(G,s,U,\frac{\varepsilon}{4t_{\max}},\frac{1}{n_{\max}^{3}}) to obtain D,RUD,R\subseteq U and disjoint sets S^v{\hat{S}}_{v} (where D=vRS^vUD=\bigcup_{v\in R}{\hat{S}}_{v}\cap U)
7:for each vRv\in R do
8:  Let GvG_{v} be the graph with vertices VS^vV\setminus{\hat{S}}_{v} contracted to a single vertex xvx_{v}
9:  Add edges with weight Lap\MT@delim@Auto\paren@star8tmaxε\text{Lap}\MT@delim@Auto\paren@star{\frac{8t_{\max}}{\varepsilon}} from xvx_{v} to every other vertex in GvG_{v}, truncating resulting edge weights to be at least 0
10:  UvS^vUU_{v}\leftarrow{\hat{S}}_{v}\cap U
11:  If |Uv|>1|U_{v}|>1, recursively set (Tv,fv)PrivateGHSteinerTree(Gv,Uv,ε,t+1,nmax)(T_{v},f_{v})\leftarrow{\color[rgb]{0.8,0.25,0.33}{\text{PrivateGHSteinerTree}{}(G_{v},U_{v},\varepsilon,t+1,n_{\max})}}
12:end for
13:Let GlargeG_{\text{large}} be the graph GG with (disjoint) vertex sets S^v{\hat{S}}_{v} contracted to single vertices yvy_{v} for all vDv\in D
14:UlargeUDU_{\text{large}}\leftarrow U\setminus D
15:If |Ularge|>1|U_{\text{large}}|>1, recursively set (Tlarge,flarge)PrivateGHSteinerTree(Glarge,Ularge,ε,t+1,nmax)(T_{\text{large}},f_{\text{large}})\leftarrow{\color[rgb]{0.8,0.25,0.33}{\text{PrivateGHSteinerTree}{}(G_{\text{large}},U_{\text{large}},\varepsilon,t+1,n_{\max})}}
16:return Combine((Tlarge,flarge),{(Tv,fv):vR},{w(S^v):vR})((T_{\text{large}},f_{\text{large}}),\{(T_{v},f_{v}):v\in R\},\{w({\hat{S}}_{v}):v\in R\}) // The weights used in this step are not private
Algorithm 5 Combine((Tlarge,flarge),{(Tv,fv):vR},{w(S^v):vR)}((T_{\text{large}},f_{\text{large}}),\{(T_{v},f_{v}):v\in R\},\{w({\hat{S}}_{v}):v\in R)\}
1:Construct TT by starting with the disjoint union TlargevRTvT_{\text{large}}\cup\bigcup_{v\in R}T_{v} and for each vRv\in R, adding an edge between fv(xv)Uvf_{v}(x_{v})\in U_{v} and flarge(yv)Ulargef_{\text{large}}(y_{v})\in U_{\text{large}} with weight w(S^v)w({\hat{S}}_{v})
2:Construct f:VU=UlargevRUvf:V\to U=U_{\text{large}}\cup\bigcup_{v\in R}U_{v} by f(v)=flarge(v)f(v^{\prime})=f_{\text{large}}(v^{\prime}) if vVvRS^vv^{\prime}\in V\setminus\bigcup_{v\in R}{\hat{S}}_{v} and f(v)=fv(v)f(v^{\prime})=f_{v}(v^{\prime}) if vS^vv^{\prime}\in{\hat{S}}_{v} for some vRv\in R

In this section, we present the algorithm PrivateGHTree (Algorithm 3) for constructing an ε\varepsilon-DP approximate Gomory-Hu tree and analyze its approximation error and privacy guarantees. The steps that differ meaningfully from the non-private version developed in [Li21] are in color. As in [Li21], we construct the slightly more general structure of an Gomory-Hu Steiner tree as an intermediate step in Algorithm 4.

Definition 5.1.

Let G=(V,E,w)G=(V,E,w) be a weighted graph and UU a set of terminals. A Γ\Gamma-approximate Gomory-Hu Steiner tree is a weighted spanning tree TT on UU with a function f:VUf:V\to U such that f|Uf|_{U} is the identity and

  • for all distinct s,tUs,t\in U, if (u,v)(u,v) is the minimum weight edge on the unique path between ss and tt, in TT, and if UU^{\prime} is the connected component of T{(u,v)}T\setminus\{(u,v)\} containing ss, then f1(U)f^{-1}(U^{\prime}) is a Γ\Gamma-approximate Min-ss-tt-Cut with λG(s,t)wT(u,v)=wG(f1(U))λG(s,t)+Γ\lambda_{G}(s,t)\leq w_{T}(u,v)=w_{G}(f^{-1}(U^{\prime}))\leq\lambda_{G}(s,t)+\Gamma.

To construct the final approximate Gomory-Hu tree, we make a call to PrivateGHSteinerTree (Algorithm 4) with U=VU=V, the entire vertex set. The algorithm PrivateGHSteinerTree is a private version of the GHTree algorithm in [Li21]. It computes several (approximate) min cuts from a randomly sampled vertex sUs\in U by making a call to PrivateGHTreeStep (Algorithm 2) to obtain D,RUD,R\subseteq U and disjoint sets S^v{\hat{S}}_{v} (where D=vRS^vUD=\bigcup_{v\in R}{\hat{S}}_{v}\cap U). For each of these cuts S^v{\hat{S}}_{v} it constructs recursive sub-instances (Gv,Uv)(G_{v},U_{v}) where GvG_{v} is obtained by contracting VS^vV\setminus{\hat{S}}_{v} to a single vertex xvx_{v} and UvS^vUU_{v}\leftarrow{\hat{S}}_{v}\cap U. Moreover, it creates a sub-instance (Glarge,Ularge)(G_{\text{large}},U_{\text{large}}) by contracting each of S^v{\hat{S}}_{v} to a single vertex yvy_{v} for yDy\in D and setting Ularge=UDU_{\text{large}}=U\leftarrow D.

Notably, on 9, where the algorithm recurses on the graph GvG_{v} with VS^vV\setminus{\hat{S}}_{v} contracted to a single vertex xvx_{v}, we add noisy edges from xvx_{v} to all other vertices of the graph. This ensures the privacy of any actual edge from xvx_{v} in the entire recursive subtree of that instance without incurring too much error. This will imply that for any edge and any instance during the recursion, there is at most one sub-instance where the edge does not receive this privacy guarantee. If tt is the depth of the recursion tree, this allows us to apply basic composition over only O(t)O(t) computations of the algorithm. Essentially, there is only one path down the recursion tree on which we need to track privacy for any given edge in the original graph. We enforce t<tmaxt<t_{\max}, and as we will show, the algorithm successfully terminates with depth less than tmaxt_{\max} with high probability.

To combine the solutions to the recursive sub-problems, we use the Combine algorithm (Algorithm 5) from [Li21], which in turn is similar to the original Gomory-Hu combine step except that it combines more than two recursive sub-instances.

Finally, Algorithm 3 calls Algorithm 4 with privacy budget ε/2\varepsilon/2. To be able to output weights of the tree edges, it simply adds Laplace noise Lap(12(n1)ε)\text{Lap}(\frac{1}{2(n-1)\varepsilon}) to each of them, hence incurring error O(nlgnε)O(\frac{n\lg n}{\varepsilon}) with high probability. This also has privacy loss ε/2\varepsilon/2 by basic composition, so the full algorithm is ε\varepsilon-differentially private.

5.1 Correctness

In this section, we analyze the approximation guarantee of our algorithm. The main lemma states that Algorithm 4 outputs an O(npolylog(n))O(n\operatorname{polylog}(n))-approximate Gomory-Hu Steiner tree.

Lemma 5.1.

Let tmax=Clg2nt_{\max}=C\lg^{2}n for a sufficiently large constant CC. PrivateGHSteinerTree(G,V,ε,0,n)(G,V,\varepsilon,0,n) outputs an O(nlg8nε)O(\frac{n\lg^{8}n}{\varepsilon})-approximate Gomory-Hu Steiner tree TT of GG with high probability.

We start by proving a lemma for analyzing a single recursive step of the algorithm. It is similar to [Li21, Lemma 4.5.4] but its proof requires a more careful application of the submodularity lemma.

Lemma 5.2.

With high probability, for any distinct vertices p,qUlargep,q\in U_{\text{large}}, we have that λG(p,q)λGlarge(p,q)λG(p,q)+O(nlg5nε)\lambda_{G}(p,q)\leq\lambda_{G_{\text{large}}}(p,q)\leq\lambda_{G}(p,q)+O(\frac{n\lg^{5}n}{\varepsilon}). Also with high probability, for any vRv\in R and distinct vertices p,qUvp,q\in U_{v}, we have that λG(p,q)λGv(p,q)λG(p,q)+O(nlg6nε)\lambda_{G}(p,q)\leq\lambda_{G_{v}}(p,q)\leq\lambda_{G}(p,q)+O(\frac{n\lg^{6}n}{\varepsilon}).

Proof.

Let us start by upper bounding how close the cuts S^v{\hat{S}}_{v} are to being Min-ss-vv-Cuts and how close the approximate min cut values w^(S^v){\hat{w}}({\hat{S}}_{v}) are to the true sizes w(S^v)w({\hat{S}}_{v}). Algorithm 4 calls Algorithm 2 with privacy parameter ε1=ε4tmax=Θ\MT@delim@Auto\paren@starεlg2n\varepsilon_{1}=\frac{\varepsilon}{4t_{\max}}=\Theta\MT@delim@Auto\paren@star{\frac{\varepsilon}{\lg^{2}n}}. By Lemma 4.1 with privacy ε1\varepsilon_{1}, it follows that (a) the sets {S^v}\{{\hat{S}}_{v}\} are approximate minimum isolating cuts with total error O\MT@delim@Auto\paren@starnlg5nεO\MT@delim@Auto\paren@star{\frac{n\lg^{5}n}{\varepsilon}} and (b) each set S^v{\hat{S}}_{v} is an approximate Min-ss-vv-Cut with error O\MT@delim@Auto\paren@starnlg6nεO\MT@delim@Auto\paren@star{\frac{n\lg^{6}n}{\varepsilon}}. On any given call to Algorithm 2, these error bounds hold with probability 1nmax31-n_{\max}^{-3} where nmaxn_{\max} is the number of vertices in the original graph. As each call to Algorithm 2 ultimately contributes an edge to the final Gomory-Hu tree via Algorithm 5, there can be at most nmax1n_{\max}-1 calls throughout the entire recursion tree, resulting in failure probability of nmax2n_{\max}^{-2} overall after a union bound.

The fact that λG(p,q)λGlarge(p,q)\lambda_{G}(p,q)\leq\lambda_{G_{\text{large}}}(p,q) follows since GlargeG_{\text{large}} is a contraction of GG. To prove the second inequality, let SS be one side of the true Min-pp-qq-Cut. Let R1=RSR_{1}=R\cap S and R2=RSR_{2}=R\setminus S. We show that the cut S=(SvR1S^v)(vR2S^v)S^{*}=(S\cup\bigcup_{v\in R_{1}}{\hat{S}}_{v})\setminus(\bigcup_{v\in R_{2}}{\hat{S}}_{v}) is an O(nlg6n/ε)O(n\lg^{6}n/\varepsilon)-approximate min cut. Since SS^{*} is also a cut in GlargeG_{\text{large}}, the desired bound on λGlarge(p,q)\lambda_{G_{\text{large}}}(p,q) follows.

Let v1,,v|R1|v_{1},\dots,v_{|R_{1}|} be the vertices of R1R_{1} in an arbitrary order. By |R1||R_{1}| applications of the submodularity lemma (Lemma 2.1),

w(SvR1S^v)w(S)+i=1|R1|(w(S^vi)w((Sj<iS^vj)S^vi)).w\left(S\cup\bigcup_{v\in R_{1}}{\hat{S}}_{v}\right)\leq w(S)+\sum_{i=1}^{|R_{1}|}\left(w({\hat{S}}_{v_{i}})-w\left(\left(S\cup\bigcup_{j<i}{\hat{S}}_{v_{j}}\right)\cap{\hat{S}}_{v_{i}}\right)\right).

Note that SvR1S^vS\cup\bigcup_{v\in R_{1}}{\hat{S}}_{v} is still a (p,q)(p,q)-cut as p,qUlargep,q\in U_{\text{large}} and the sets S^vi{\hat{S}}_{v_{i}} are each disjoint from UlargeU_{\text{large}}. Moreover, for each ii, SS contains viv_{i} and so (Sj<iS^vj)S^vi\left(S\cup\bigcup_{j<i}{\hat{S}}_{v_{j}}\right)\cap{\hat{S}}_{v_{i}} isolates viv_{i} from all vertices in VS^viV\setminus{\hat{S}}_{v_{i}}. The S^vi{\hat{S}}_{v_{i}} are the outputs of the private isolating cuts algorithm (Algorithm 1). Using the fact that {S^v}\{{\hat{S}}_{v}\} are approximate minimum isolating cuts, the sum in the RHS above can be upper bounded by O(nlg5nε)O(\frac{n\lg^{5}n}{\varepsilon}). Letting S=SvR1S^vS^{\prime}=S\cup\bigcup_{v\in R_{1}}{\hat{S}}_{v}, and S′′=(VS)vR2S^vS^{\prime\prime}=(V\setminus S^{\prime})\cup\bigcup_{v\in R_{2}}{\hat{S}}_{v} a similar argument but applied to VSV\setminus S^{\prime}, shows that

w(S′′)=w((VS)vR2S^v)w(VS)+O\MT@delim@Auto\paren@starnlg5nε.w(S^{\prime\prime})=w\left((V\setminus S^{\prime})\cup\bigcup_{v\in R_{2}}{\hat{S}}_{v}\right)\leq w(V\setminus S^{\prime})+O\MT@delim@Auto\paren@star{\frac{n\lg^{5}n}{\varepsilon}}.

But S=VS′′S^{*}=V\setminus S^{\prime\prime}, so we get that

w(S)=w(S′′)w(S)+O\MT@delim@Auto\paren@starnlg5nεw(S)+O\MT@delim@Auto\paren@starnlg5nε,w(S^{*})=w(S^{\prime\prime})\leq w(S^{\prime})+O\MT@delim@Auto\paren@star{\frac{n\lg^{5}n}{\varepsilon}}\leq w(S)+O\MT@delim@Auto\paren@star{\frac{n\lg^{5}n}{\varepsilon}},

as desired.

For the case of p,qUvp,q\in U_{v} for some vv, again the bound λG(p,q)λGu(p,q)\lambda_{G}(p,q)\leq\lambda_{G_{u}}(p,q) is clear. Thus, it suffices to consider the upper bound on λGv(p,q)\lambda_{G_{v}}(p,q). Let SS be the side of a Min-pp-qq-Cut in GG which does not contain vv. Assume first that sSs\notin S. By the submodularity lemma (Lemma 2.1)

w(SS^v)w(S)+w(S^v)w(SS^v).w(S\cap{\hat{S}}_{v})\leq w(S)+w({\hat{S}}_{v})-w(S\cup{\hat{S}}_{v}).

By the approximation guarantees of Algorithm 2, w(S^v)λG(s,v)+O(nlg6nε)w({\hat{S}}_{v})\leq\lambda_{G}(s,v)+O(\frac{n\lg^{6}n}{\varepsilon}). Moreover, note that SS^vS\cup{\hat{S}}_{v} is an (s,v)(s,v)-cut of GG, so w(SS^v)λG(s,v)w(S\cup{\hat{S}}_{v})\geq\lambda_{G}(s,v). Thus,

w(SS^v)w(S)+O\MT@delim@Auto\paren@starnlg6nε=λG(p,q)+O\MT@delim@Auto\paren@starnlg6nε.w(S\cap{\hat{S}}_{v})\leq w(S)+O\MT@delim@Auto\paren@star{\frac{n\lg^{6}n}{\varepsilon}}=\lambda_{G}(p,q)+O\MT@delim@Auto\paren@star{\frac{n\lg^{6}n}{\varepsilon}}.

Since SS^vS\cap{\hat{S}}_{v} is a (p,q)(p,q)-cut of GvG_{v}, we must have that w(SS^v)λGv(p,q)w(S\cap{\hat{S}}_{v})\geq\lambda_{G_{v}}(p,q), so in conclusion λGv(p,q)λG(p,q)+O(nlg6nε)\lambda_{G_{v}}(p,q)\leq\lambda_{G}(p,q)+O(\frac{n\lg^{6}n}{\varepsilon}) ignoring the added noisy edges to GvG_{v} in 9 of Algorithm 4.

Adding the noisy edges can only increase the cost by O(nlg3nε)O(\frac{n\lg^{3}n}{\varepsilon}) with high probability via Laplace tail bounds (note that there can be at most n1n-1 noisy edges added as each time a noisy edge is added an edge is added to the final approximate Gomory Hu Steiner tree). This finishes the proof in the case sSs\notin S. A similar argument handles the case where sSs\in S but here we relate the value w(VS)w(V\setminus S) to (w(VS)S^v)(w(V\setminus S)\cap{\hat{S}}_{v}). \Box

To bound the error of the algorithm, we need a further lemma bounding the depth of its recursion. The argument is similar to that of [Li21].

Lemma 5.3.

If tmax=Clg2nt_{\max}=C\lg^{2}n for a sufficiently large constant CC, then, with high probability, no recursive call to Algorithm 4 from PrivateGHTree(G,ε)(G,\varepsilon) aborts.

Proof.

Each of the recursive instances, (Gv,Uv)(G_{v},U_{v}) has |Uv|910|U||U_{v}|\leq\frac{9}{10}|U| by the way they are picked in 7 of Algorithm 2. Moreover, by a lemma from [AKT20], if ss is picked uniformly at random from UU, then 𝔼[D]=Ω(|U|1){\mathbb{E}}\left[D^{*}\right]=\Omega(|U|-1). By Lemma 4.1, the expected size of vRS^v\bigcup_{v\in R}{\hat{S}}_{v} returned by a call to Algorithm 2 when picking ss at random from |U||U| is then at least Ω\MT@delim@Auto\paren@star|U|1lgn\Omega\MT@delim@Auto\paren@star{\frac{|U|-1}{\lg n}} with constant probability. It follows that, 𝔼[|Ularge|]|U|(1Ω(1/lgn)){\mathbb{E}}\left[|U_{\text{large}}|\right]\leq|U|(1-\Omega(1/\lg n)) when |U|1|U|\geq 1. By Markov’s inequality and union bound, all sub-instances have |U|=1|U|=1 within O(lg2n)O(\lg^{2}n) recursive depth with high probability. \Box

We can now prove Lemma 5.1. The argument is again similar to [Li21] except we have to incorporate the approximation errors.

Proof of Lemma 5.1.

By Lemma 5.3, the algorithm does not abort with high probability.

Let Δ=O(nlg6nε)\Delta=O(\frac{n\lg^{6}n}{\varepsilon}) be such that with high probability λGv(p,q)λG(p,q)+Δ\lambda_{G_{v}}(p,q)\leq\lambda_{G}(p,q)+\Delta for p,qUvp,q\in U_{v} and similarly λGlarge(p,q)λG(p,q)+Δ\lambda_{G_{large}}(p,q)\leq\lambda_{G}(p,q)+\Delta for p,qUlargep,q\in U_{large}. The existence of Δ\Delta is guaranteed by Lemma 5.2. We prove by induction on i=0,,tmaxi=0,\dots,t_{\max}, that the output to the instances at level tmaxit_{\max}-i of the recursion are 2iΔ2i\Delta-approximate Gomory-Hu Steiner trees. This holds trivially for i=0i=0 as the instances on that level have |U|=1|U|=1 and the tree is the trivial one-vertex tree approximating no cuts at all. Let i1i\geq 1 and assume inductively that the result holds for smaller ii. In particular, if (T,f)(T,f) is the output of an instance at recursion level ii, then the trees (Tv,fv)(T_{v},f_{v}) and (Tlarge,flarge)(T_{\text{large}},f_{\text{large}}) are 2(i1)Δ2(i-1)\Delta-approximate Gomory-Hu Steiner trees of their respective GvG_{v} or GlargeG_{\text{large}} graphs.

Consider any internal edge (a,b)Tlarge(a,b)\in T_{\text{large}} (without loss of generality, what follows also holds for (a,b)Tv)(a,b)\in T_{v}). Let UU^{\prime} and UlargeU^{\prime}_{\text{large}} be the connected component containing aa after removing (a,b)(a,b) from TT and TlargeT_{\text{large}}, respectively. By design of Algorithm 5, flarge1(Ularge)f^{-1}_{\text{large}}(U^{\prime}_{\text{large}}) and f1(U)f^{-1}(U^{\prime}) are the same except each contracted vertex yvflarge1(Ularge)y_{v}\in f^{-1}_{\text{large}}(U^{\prime}_{\text{large}}) appears as S^vf1(U){\hat{S}}_{v}\subseteq f^{-1}(U^{\prime}). It follows that wGlarge(flarge1(Ularge))=wG(f1(U))w_{G_{\text{large}}}(f_{\text{large}}^{-1}(U^{\prime}_{\text{large}}))=w_{G}(f^{-1}(U^{\prime})). By the inductive hypothesis, (Tlarge,flarge)(T_{\text{large}},f_{\text{large}}) is an approximate Gomory-Hu Steiner tree, so wGlarge(flarge1(Ularge))=wTlarge(a,b)w_{G_{\text{large}}}(f_{\text{large}}^{-1}(U^{\prime}_{\text{large}}))=w_{T_{\text{large}}}(a,b). Therefore setting wT(a,b)=wTlarge(a,b)=wG(f1(U))w_{T}(a,b)=w_{T_{\text{large}}}(a,b)=w_{G}(f^{-1}(U^{\prime})) has the correct cost for TT according to the definition of an approximate Gomory-Hu Steiner tree.

Furthermore, on the new edges (fv(xv),flarge(yv))(f_{v}(x_{v}),f_{\text{large}}(y_{v})), the weight w(S^v)w({\hat{S}}_{v}) is the correct weight for that edge in TT as S^v{\hat{S}}_{v} is the fv(xv)f_{v}(x_{v}) side of the connected component after removing that edge. Finally, by adding these new edges, the resulting tree is a spanning tree. So, the structure of the tree is correct, and it remains to argue that the cuts induced by the tree (via minimum edges on shortest paths) are approximate Min-ss-tt-Cuts.

Consider any p,qUp,q\in U. Let (a,b)(a,b) be the minimum edge on the shortest path in TT. Note that it is always the case that wT(a,b)λ(a,b)w_{T}(a,b)\geq\lambda(a,b) as wT(a,b)w_{T}(a,b) corresponds to the value of a cut in GG separating aa and bb. We will proceed by cases.

If (a,b)Tlarge(a,b)\in T_{\text{large}}, then by induction, wT(a,b)=wTlarge(a,b)λGlarge(a,b)+(2i2)Δw_{T}(a,b)=w_{T_{\text{large}}}(a,b)\leq\lambda_{G_{\text{large}}}(a,b)+(2i-2)\Delta. By Lemma 5.2, it follows that wT(a,b)λG(a,b)+(2i1)Δw_{T}(a,b)\leq\lambda_{G}(a,b)+(2i-1)\Delta. The exact same argument applies if (a,b)Tv(a,b)\in T_{v} for some vRv\in R.

The case that remains is if (a,b)(a,b) is a new edge with a=fv1(xv)Uva=f_{v}^{-1}(x_{v})\in U_{v} and b=flarge1(yv)Ulargeb=f_{\text{large}}^{-1}(y_{v})\in U_{\text{large}}. Then, wT(a,b)=wG(S^v)w_{T}(a,b)=w_{G}({\hat{S}}_{v}). By Lemma 4.1, S^v{\hat{S}}_{v} is an approximate Min-vv-ss-Cut and wT(a,b)λG(v,s)+Δw_{T}(a,b)\leq\lambda_{G}(v,s)+\Delta. To connect this value to λG(a,b)\lambda_{G}(a,b), note that over the choices of where vv and ss lie on the Min-aa-bb-Cut,

λG(a,b)min\MT@delim@Auto\brace@starλG(a,v),λG(v,s),λG(s,b).\lambda_{G}(a,b)\geq\min\MT@delim@Auto\brace@star{\lambda_{G}(a,v),\lambda_{G}(v,s),\lambda_{G}(s,b)}.

Let SaS^{\prime}_{a} be the aa side of the (a,v)(a,v) cut induced by the approximate Gomory-Hu Steiner tree (Tv,fv)(T_{v},f_{v}). As a=fv1(xv)a=f_{v}^{-1}(x_{v}) and sxvs\in x_{v}, SaS^{\prime}_{a} is also the ss side of a (v,s)(v,s) cut. Therefore, wG(Sa)λG(v,s)w_{G}(S^{\prime}_{a})\geq\lambda_{G}(v,s). On the other hand, by our inductive hypothesis and Lemma 5.2, this is an approximate Min-aa-vv-Cut: wG(Sa)λG(a,v)+(2i1)Δw_{G}(S^{\prime}_{a})\leq\lambda_{G}(a,v)+(2i-1)\Delta. The analogous argument holds to show λG(v,s)λG(s,b)+(2i1)Δ\lambda_{G}(v,s)\leq\lambda_{G}(s,b)+(2i-1)\Delta. Therefore,

λG(v,s)min\MT@delim@Auto\brace@starλG(a,v),λG(s,b)+(2i1)Δ\lambda_{G}(v,s)\leq\min\MT@delim@Auto\brace@star{\lambda_{G}(a,v),\lambda_{G}(s,b)}+(2i-1)\Delta

and

wT(a,b)λG(v,s)+ΔλG(a,b)+2iΔ.w_{T}(a,b)\leq\lambda_{G}(v,s)+\Delta\leq\lambda_{G}(a,b)+2i\Delta.

In all cases, wT(a,b)λG(a,b)+2iΔw_{T}(a,b)\leq\lambda_{G}(a,b)+2i\Delta. As the cut corresponding to the edge (a,b)(a,b) is on the path from pp to qq, it is also a (p,q)(p,q)-cut, so λG(p,q)wG(a,b)\lambda_{G}(p,q)\leq w_{G}(a,b). Furthermore, it must the case that there is an edge (a,b)(a^{\prime},b^{\prime}) along the path between pp to qq such that aa^{\prime} and bb^{\prime} are in different sides of the true Min-pp-qq-Cut. Otherwise, pp and qq would be on the same side of the cut. Therefore, λG(p,q)λG(a,b)\lambda_{G}(p,q)\geq\lambda_{G}(a^{\prime},b^{\prime}). As we chose (a,b)(a,b) to be the minimum weight edge,

wT(a,b)wT(a,b)λG(a,b)+2iΔλG(p,q)+2iΔ.w_{T}(a,b)\leq w_{T}(a^{\prime},b^{\prime})\leq\lambda_{G}(a^{\prime},b^{\prime})+2i\Delta\leq\lambda_{G}(p,q)+2i\Delta.

This completes the induction. It follows that the call to PrivateGHSteinerTree(G,V,ε,0)\text{PrivateGHSteinerTree}{}(G,V,\varepsilon,0) outputs a 2tmaxΔ2t_{\max}\Delta-approximate Gomory-Hu Steiner tree TT. Substituting in the values tmax=O(lg2n)t_{\max}=O(\lg^{2}n) and Δ=O(nlg6nε)\Delta=O(\frac{n\lg^{6}n}{\varepsilon}) gives the approximation guarantee. \Box

We now state our main result on the approximation guarantee of Algorithm 3.

Theorem 5.1.

Let T=(VT,ET,wT)T=(V_{T},E_{T},w_{T}) be the weighted tree output by PrivateGHTree(G=(V,E,w),ε)(G=(V,E,w),\varepsilon) on a weighted graph GG. For each edge eETe\in E_{T}, define SeS_{e} to be the set of vertices of one of the connected components of T{e}T\setminus\{e\}. Let u,vVu,v\in V be distinct vertices and let emine_{\min} be an edge on the unique uu-vv path in TT such that wT(emin)w_{T}(e_{\min}) is minimal. With high probability, SeminS_{e_{\min}} is an O(nlg8nε)O(\frac{n\lg^{8}n}{\varepsilon})-approximate Min-uu-vv-Cut and moreover, |λG(u,v)wT(emin)|=O(nlg8nε)|\lambda_{G}(u,v)-w_{T}(e_{\min})|=O(\frac{n\lg^{8}n}{\varepsilon}).

Proof.

Note that for each edge ee, wT(e)w_{T}(e) is obtained by adding noise Lap\MT@delim@Auto\paren@star2(n1)ε\text{Lap}\MT@delim@Auto\paren@star{\frac{2(n-1)}{\varepsilon}} to the cut value w(Se)w(S_{e}). Thus, |w(Se)wT(e)|=O(nlgnε)|w(S_{e})-w_{T}(e)|=O(\frac{n\lg n}{\varepsilon}) with high probability for all eTe\in T. Now let e0e_{0} be an edge on the unique uu-vv path in TT such that w(Se0)w(S_{e_{0}}) is minimal. Then, by Lemma 5.1, w(Se0)λG(u,v)+O(nlg8nε)w(S_{e_{0}})\leq\lambda_{G}(u,v)+O(\frac{n\lg^{8}n}{\varepsilon}) with high probability. As emine_{\min} was chosen as an edge on the uu-vv path in TT of minimal weight, wT(emin)wT(e0)w_{T}(e_{\min})\leq w_{T}(e_{0}), and so

w(Semin)\displaystyle w(S_{e_{\min}}) wT(emin)+O\MT@delim@Auto\paren@starnlgnεwT(e0)+O\MT@delim@Auto\paren@starnlgnε\displaystyle\leq w_{T}(e_{\min})+O\MT@delim@Auto\paren@star{\frac{n\lg n}{\varepsilon}}\leq w_{T}(e_{0})+O\MT@delim@Auto\paren@star{\frac{n\lg n}{\varepsilon}}
w(Se0)+O\MT@delim@Auto\paren@starnlgnελG(u,v)+O\MT@delim@Auto\paren@starnlg8nε.\displaystyle\leq w(S_{e_{0}})+O\MT@delim@Auto\paren@star{\frac{n\lg n}{\varepsilon}}\leq\lambda_{G}(u,v)+O\MT@delim@Auto\paren@star{\frac{n\lg^{8}n}{\varepsilon}}.

On the the other hand, SS defines a (u,v)(u,v)-cut, so λG(u,v)w(S)\lambda_{G}(u,v)\leq w(S). This proves the first statement. Moreover, the string of inequalities above combined with λG(u,v)w(S)\lambda_{G}(u,v)\leq w(S) in particular entails that

λG(u,v)wT(emin)+O\MT@delim@Auto\paren@starnlgnελG(u,v)+O\MT@delim@Auto\paren@starnlg8nε,\lambda_{G}(u,v)\leq w_{T}(e_{\min})+O\MT@delim@Auto\paren@star{\frac{n\lg n}{\varepsilon}}\leq\lambda_{G}(u,v)+O\MT@delim@Auto\paren@star{\frac{n\lg^{8}n}{\varepsilon}},

from which the final statement follows. \Box

5.2 Privacy

We will refer to any subroutine or algorithm as being ε\varepsilon-DP(G,G)(G,G^{\prime}) if it satisfies the ε\varepsilon-DP condition for a fixed pair of neighboring graphs G,GG,G^{\prime}. We will prove privacy by proving that ε\varepsilon-DP(G,G)(G,G^{\prime}) holds for any G,GG,G^{\prime}.

Theorem 5.2.

PrivateGHTree(G,U,ε)(G,U,\varepsilon) is ε\varepsilon-DP.

Proof.

Let σGH(t)\sigma_{GH}(t) be the privacy of the topology of the tree released by PrivateGHSteinerTree on a graph with at most nn vertices, privacy parameter ε\varepsilon, and at recursive depth tt. Note that the edge weights of the tree are not private but no computation depends on these values. Until the end of the analysis, we will ignore the edge weights on the tree in terms of privacy. We want to show σGH(0)ε\sigma_{GH}(0)\leq\varepsilon.

Consider two neighboring graphs GG and GG^{\prime} which differ by 11 on the weight on a single edge e=(a,b)e=(a,b). Consider a call to PrivateGHSteinerTree with privacy parameter ε\varepsilon and recursion depth tt. Let za,zbz_{a},z_{b} be the vertices containing a,ba,b (possibly after contractions). If za=zbz_{a}=z_{b}, the output of PrivateGHSteinerTree is 0-DP(G,G)(G,G^{\prime}) as ee has been contracted and will have no effect on the output of the algorithm. By Lemma 4.3, the call to PrivateGHTreeStep is ε4tmax\frac{\varepsilon}{4t_{\max}}-DP. Manipulation of the resulting vertex sets S^v{\hat{S}}_{v} does not hurt privacy via post-processing and as they do not depend on edge structure or weights. We will proceed by cases on where za,zbz_{a},z_{b} belong.

Case 1: za,zbS^vz_{a},z_{b}\in{\hat{S}}_{v}

In this setting, za,zbz_{a},z_{b} belong to the same approximate min isolating cut. From the for loop on 7, releasing any (Tv,fv)(T_{v^{\prime}},f_{v^{\prime}}) for vvv^{\prime}\neq v is 0-DP(G,G)(G,G^{\prime}) as za,zbz_{a},z_{b} will be contracted to the same vertex xvx_{v^{\prime}}. On the other hand, za,zbz_{a},z_{b} will exist as single vertices in GvG_{v} and releasing (Tv,fv)(T_{v^{\prime}},f_{v^{\prime}}) is σGH(t+1)\sigma_{GH}(t+1)-DP(G,G)(G,G^{\prime}). Finally, releasing (Tlarge,flarge)(T_{\text{large}},f_{\text{large}}) is 0-DP(G,G)(G,G^{\prime}) as za,zbz_{a},z_{b} will be both contracted to yvy_{v}. The Combine algorithm is only doing post-processing and does not hurt privacy. By basic composition (Theorem 2.1), the overall privacy in this case is

ε4tmax+σGH(t+1).\frac{\varepsilon}{4t_{\max}}+\sigma_{GH}(t+1).
Case 2: zaS^v,zbS^vz_{a}\in{\hat{S}}_{v},z_{b}\in{\hat{S}}_{v^{\prime}}

Here, za,zbz_{a},z_{b} belong to separate approximate min isolating cuts S^v,S^v{\hat{S}}_{v},{\hat{S}}_{v^{\prime}}. For any v′′{v,v}v^{\prime\prime}\notin\{v,v^{\prime}\}, releasing (Tv′′,fv′′)(T_{v^{\prime\prime}},f_{v^{\prime\prime}}) is 0-DP(G,G)(G,G^{\prime}). Consider releasing (Tv,fv)(T_{v},f_{v}) where zaz_{a} will be included in GvG_{v} and zbz_{b} will be contracted to xvx_{v}. On 9, we add Lap\MT@delim@Auto\paren@star8tmaxε\text{Lap}\MT@delim@Auto\paren@star{\frac{8t_{\max}}{\varepsilon}} to each edge between xvx_{v} and all vertices in S^v{\hat{S}}_{v}, including to ee, and as a post-processing step, we truncate the edge weights to zero. This ensures that releasing (Tv,fv)(T_{v},f_{v}) is ε8tmax\frac{\varepsilon}{8t_{\max}}-DP(G,G)(G,G^{\prime}) via the Laplace mechanism (Theorem 2.2). Any computation done in the resulting recursive branch does not hurt privacy via post-processing. The same argument applies without loss of generality to releasing (Tv,fv)(T_{v^{\prime}},f_{v^{\prime}}). Releasing (Tlarge,flarge)(T_{\text{large}},f_{\text{large}}) involves a recursive call to a graph where za,zbz_{a},z_{b} are contracted to separate vertices yv,yvy_{v},y_{v^{\prime}}, and is σGH(t+1)\sigma_{GH}(t+1) private. The overall privacy in this case is

ε4tmax+2ε8tmax+σGH(t+1)=ε2tmax+σGH(t+1).\frac{\varepsilon}{4t_{\max}}+2\frac{\varepsilon}{8t_{\max}}+\sigma_{GH}(t+1)=\frac{\varepsilon}{2t_{\max}}+\sigma_{GH}(t+1).
Case 3: zaS^v,zbVvRS^vz_{a}\in{\hat{S}}_{v},z_{b}\in V\setminus\bigcup_{v\in R}{\hat{S}}_{v}

Here, zaz_{a} belongs to an approximate min isolating cut and zbz_{b} does not. Once again, for any vvv^{\prime}\neq v, releasing (Tv,fv)(T_{v^{\prime}},f_{v^{\prime}}) is 0-DP(G,G)(G,G^{\prime}). By the same argument above via the Laplace mechanism, releasing (Tv,fv)(T_{v},f_{v}) is ε8tmax\frac{\varepsilon}{8t_{\max}}-DP(G,G)(G,G^{\prime}). Finally, releasing (Tlarge,flarge)(T_{\text{large}},f_{\text{large}}) is σGH(t+1)\sigma_{GH}(t+1) private. The overall privacy, in this case, is

ε4tmax+ε8tmax+σGH(t+1)=3ε8tmax+σGH(t+1)\frac{\varepsilon}{4t_{\max}}+\frac{\varepsilon}{8t_{\max}}+\sigma_{GH}(t+1)=\frac{3\varepsilon}{8t_{\max}}+\sigma_{GH}(t+1)
Case 4: za,zbVvRS^vz_{a},z_{b}\in V\setminus\bigcup_{v\in R}{\hat{S}}_{v}

In this case, neither zaz_{a} nor zbz_{b} belong to an approximate min isolating cut. Therefore, releasing all (Tv,fv)(T_{v^{\prime}},f_{v^{\prime}}) is 0-DP(G,G)(G,G^{\prime}). Releasing (Tlarge,flarge)(T_{\text{large}},f_{\text{large}}) is σGH(t+1)\sigma_{GH}(t+1) private. The overall privacy, in this case, is

ε4tmax+σGH(t+1)\frac{\varepsilon}{4t_{\max}}+\sigma_{GH}(t+1)

Across all cases, the maximum privacy cost is incurred in Case 2. By basic composition, the privacy of the algorithm is bounded by the recurrence

σGH(t)ε2tmax+σGH(t+1).\sigma_{GH}(t)\leq\frac{\varepsilon}{2t_{\max}}+\sigma_{GH}(t+1).

As the recurrence ends at t<tmaxt^{\prime}<t_{\max},

σGH(0)tmax\MT@delim@Auto\paren@starε2tmax=ε2.\sigma_{GH}(0)\leq t_{\max}\MT@delim@Auto\paren@star{\frac{\varepsilon}{2t_{\max}}}=\frac{\varepsilon}{2}.

Therefore, releasing the unweighted tree from PrivateGHTree is ε2\frac{\varepsilon}{2}-DP. Each edge weight in the tree comes from calculating the weight of a given cut, which can change by at most 11 between two neighboring graphs. As the tree contains n1n-1 edges, all tree weights can be released via the Laplace mechanism by adding Lap\MT@delim@Auto\paren@star2(n1)ε\text{Lap}\MT@delim@Auto\paren@star{\frac{2(n-1)}{\varepsilon}} noise to each edge weight, resulting in ε2\frac{\varepsilon}{2}-DP. Then, releasing the weighted tree is ε\varepsilon-DP, as required. \Box

5.3 Runtime

While runtime is not our main focus, as a final note, our algorithm can be implemented to run in near-quadratic time in the number of vertices of the graph. The runtime is inherited directly from prior work of [AKL+22], which utilizes the same recursive algorithm introduced in [Li21]. The overall structure of their main algorithm and subroutines remains in our work with changes of the form (a) altering runtime-independent conditions in if statements or (b) adding noise to cut values or edges in the graph. While left unspecified here, computation of single source Min-ss-vv-Cuts in Algorithm 2 should be done via the runtime-optimized algorithm of prior work to achieve the best bound. Then, via Theorem 1.3 of [AKL+22], Algorithm 3 runs in time O~(n2)\tilde{O}(n^{2}).

References

  • [AKL+22] Amir Abboud, Robert Krauthgamer, Jason Li, Debmalya Panigrahi, Thatchaphol Saranurak, and Ohad Trabelsi. Breaking the cubic barrier for all-pairs max-flow: Gomory-Hu tree in nearly quadratic time. In Proceedings of the 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 884–895, 10 2022.
  • [AKT20] Amir Abboud, Robert Krauthgamer, and Ohad Trabelsi. Cut-equivalent trees are optimal for min-cut queries. In Proceedings of the 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), pages 105–118, 2020.
  • [AKT21] Amir Abboud, Robert Krauthgamer, and Ohad Trabelsi. Subcubic algorithms for Gomory-Hu tree in unweighted graphs. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing (STOC), pages 1725–1737, 2021.
  • [AKT22] Amir Abboud, Robert Krauthgamer, and Ohad Trabelsi. APMF << APSP? Gomory-Hu tree for unweighted graphs in almost-quadratic time. In Proceedings of the 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), pages 1135–1146, 2022.
  • [ALPS23] Amir Abboud, Jason Li, Debmalya Panigrahi, and Thatchaphol Saranurak. All-pairs max-flow is no harder than single-pair max-flow: Gomory-Hu trees in almost-linear time. In Proceedings of the 2023 IEEE 64th Annual Symposium on Foundations of Computer Science (FOCS), pages 2204–2212, 2023.
  • [AU19] Raman Arora and Jalaj Upadhyay. On differentially private graph sparsification and applications. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS), pages 13378–13389, 2019.
  • [BBDS12] Jeremiah Blocki, Avrim Blum, Anupam Datta, and Or Sheffet. The Johnson-Lindenstrauss transform itself preserves differential privacy. In Proceedings of the 53rd Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 410–419, 2012.
  • [BBDS13] Jeremiah Blocki, Avrim Blum, Anupam Datta, and Or Sheffet. Differentially private data analysis of social networks via restricted sensitivity. In Proceedings of the 13th Innovations in Theoretical Computer Science (ITCS), pages 87–96, 2013.
  • [BDG+24] Greg Bodwin, Chengyuan Deng, Jie Gao, Gary Hoppenworth, Jalaj Upadhyay, and Chen Wang. The discrepancy of shortest paths. In International Colloquium on Automata, Languages and Programming, 2024.
  • [BDK07] Lars Backstrom, Cynthia Dwork, and Jon Kleinberg. Wherefore art thou R3579X? anonymized social networks, hidden patterns, and structural steganography. In Proceedings of the 16th international conference on World Wide Web (WWW), pages 181–190, 2007.
  • [BEK21] Mark Bun, Marek Elias, and Janardhan Kulkarni. Differentially private correlation clustering. In Proceedings of the 38th International Conference on Machine Learning (ICML), pages 1136–1146, 2021.
  • [BKM+22] Amos Beimel, Haim Kaplan, Yishay Mansour, Kobbi Nissim, Thatchaphol Saranurak, and Uri Stemmer. Dynamic algorithms against an adaptive adversary: Generic constructions and lower bounds. In Proceedings of the 54th Annual ACM SIGACT Symposium on Theory of Computing (STOC), pages 1671–1684, 2022.
  • [BLR13] Avrim Blum, Katrina Ligett, and Aaron Roth. A learning theory approach to noninteractive database privacy. J. ACM, 60(2):1–25, 2013.
  • [BNSV15] Mark Bun, Kobbi Nissim, Uri Stemmer, and Salil P. Vadhan. Differentially private release and learning of threshold functions. In Proceedings of the IEEE 56th Annual Symposium on Foundations of Computer Science (FOCS), pages 634–649, 2015.
  • [BSWN15] Glencora Borradaile, Piotr Sankowski, and Christian Wulff-Nilsen. Min st-cut oracle for planar graphs with near-linear preprocessing time. ACM Trans. Algorithms, 11(3):1–29, 2015.
  • [CDFZ24] Rishi Chandra, Michael Dinitz, Chenglin Fan, and Zongrui Zou. Differentially private multiway and kk-cut, 2024.
  • [CGK+23] Justin Y. Chen, Badih Ghazi, Ravi Kumar, Pasin Manurangsi, Shyam Narayanan, Jelani Nelson, and Yinzhan Xu. Differentially private all-pairs shortest path distances: Improved algorithms and lower bounds. In Proceedings of the 2023 ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 5040–5067, 2023.
  • [CKL+22] Li Chen, Rasmus Kyng, Yang P. Liu, Richard Peng, Maximilian Probst Gutenberg, and Sushant Sachdeva. Maximum flow and minimum-cost flow in almost-linear time. In Proceedings of the 63rd IEEE Annual Symposium on Foundations of Computer Science (FOCS), pages 612–623, 2022.
  • [CLRS22] Thomas H. Cormen, Charles E. Leiserson, Ronald L. Rivest, and Clifford Stein. Introduction to algorithms. MIT press, 2022.
  • [CQ21] Chandra Chekuri and Kent Quanrud. Isolating cuts, (bi-)submodularity, and faster algorithms for connectivity. In Proceedings of the 48th International Colloquium on Automata, Languages, and Programming (ICALP), volume 198, pages 50:1–50:20, 2021.
  • [CRB+19] Dr Chris Culnane, A Rubinstein, IP Benjamin, A Teague, et al. Stop the open data bus, we want to get off. arXiv preprint arXiv:1908.05004, 2019.
  • [Cun85] William H. Cunningham. Minimum cuts, modular functions, and matroid polyhedra. Networks, 15(2):205–215, 1985.
  • [Dan51] George B. Dantzig. Application of the simplex method to a transportation problem. Activity analysis and production and allocation, 1951.
  • [DKLV23] Michael Dinitz, Satyen Kale, Silvio Lattanzi, and Sergei Vassilvitskii. Improved differentially private densest subgraph: Local and purely additive. arXiv preprint arXiv:2308.10316, 2023.
  • [DKM+06] Cynthia Dwork, Krishnaram Kenthapadi, Frank McSherry, Ilya Mironov, and Moni Naor. Our data, ourselves: Privacy via distributed noise generation. In Proceedings of the 24th Annual International Conference on the Theory and Applications of Cryptographic Techniques (EUROCRYPT), pages 486–503, 2006.
  • [DL09] Cynthia Dwork and Jing Lei. Differential privacy and robust statistics. In Proceedings of the 41st Annual ACM Symposium on Theory of Computing (STOC), pages 371–380, 2009.
  • [DLL23] Laxman Dhulipala, George Z. Li, and Quanquan C. Liu. Near-optimal differentially private k-core decomposition. arXiv preprint arXiv:2312.07706, 2023.
  • [DLR+22] Laxman Dhulipala, Quanquan C. Liu, Sofya Raskhodnikova, Jessica Shi, Julian Shun, and Shangdi Yu. Differential privacy from locally adjustable graph algorithms: k-core decomposition, low out-degree ordering, and densest subgraphs. In Proceedings of the 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 754–765, 2022.
  • [DMN23] Mina Dalirrooyfard, Slobodan Mitrović, and Yuriy Nevmyvaka. Nearly tight bounds for differentially private multiway cut. In Proceedings of the Advances in Neural Information Processing Systems 36 (NeurIPS), 2023.
  • [DMNS06] Cynthia Dwork, Frank McSherry, Kobbi Nissim, and Adam D. Smith. Calibrating noise to sensitivity in private data analysis. In Proceedings of the 3rd Theory of Cryptography Conference (TCC), volume 3876, pages 265–284, 2006.
  • [DR14] Cynthia Dwork and Aaron Roth. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci., 9(3–4):211–407, 2014.
  • [Dwo06] Cynthia Dwork. Differential privacy. In Proceedings of the 33rd International Colloquium on Automata, Languages, and Programming (ICALP), pages 1–12, 2006.
  • [EFS56] Peter Elias, Amiel Feinstein, and Claude Shannon. A note on the maximum flow through a network. IRE Trans. Inf. Theory, 2(4):117–119, 1956.
  • [EKKL20] Marek Eliáš, Michael Kapralov, Janardhan Kulkarni, and Yin Tat Lee. Differentially private release of synthetic graphs. In Proceedings of the 14th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 560–578, 2020.
  • [ELRS23] Talya Eden, Quanquan C. Liu, Sofya Raskhodnikova, and Adam Smith. Triangle counting with local edge differential privacy. In Proceedings of the 50th International Colloquium on Automata, Languages, and Programming (ICALP), 2023.
  • [FF56] Lester Randolph Ford and Delbert R. Fulkerson. Maximal flow through a network. Can. J. Math., 8:399–404, 1956.
  • [FHS22] Alireza Farhadi, MohammadTaghi Hajiaghayi, and Elaine Shi. Differentially private densest subgraph. In Proceedings of the 25th International Conference on Artificial Intelligence and Statistics (AISTATS), pages 11581–11597, 2022.
  • [FLL22] Chenglin Fan, Ping Li, and Xiaoyun Li. Private graph all-pairwise-shortest-path distance release with improved error rate. In Proceedings of the Advances in Neural Information Processing Systems 35 (NeurIPS), 2022.
  • [GH61] Ralph E. Gomory and Tien Chung Hu. Multi-terminal network flows. J. Soc. Indust. Appl. Math., 9(4):551–570, 1961.
  • [GLM+10] Anupam Gupta, Katrina Ligett, Frank McSherry, Aaron Roth, and Kunal Talwar. Differentially private combinatorial optimization. In Proceedings of the 21st Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), 2010.
  • [GRU12] Anupam Gupta, Aaron Roth, and Jonathan R. Ullman. Iterative constructions and private data release. In Proceedings of the 9th Theory of Cryptography Conference (TCC), volume 7194, pages 339–356, 2012.
  • [HKM+20] Avinatan Hassidim, Haim Kaplan, Yishay Mansour, Yossi Matias, and Uri Stemmer. Adversarially robust streaming algorithms via differential privacy. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS), 2020.
  • [HKPB07] Ramesh Hariharan, Telikepalli Kavitha, Debmalya Panigrahi, and Anand Bhalgat. An O~(mn)\tilde{O}(mn) Gomory-Hu tree construction algorithm for unweighted graphs. In Proceedings of the 39th Annual ACM Symposium on Theory of Computing (STOC), pages 605–614, 2007.
  • [HLMJ09] Michael Hay, Chao Li, Gerome Miklau, and David D. Jensen. Accurate estimation of the degree distribution of private networks. In Proceedings of the 9th IEEE International Conference on Data Mining (ICDM), pages 169–178, 2009.
  • [HR55] T. E. Harris and F. S. Ross. Fundamentals of a method for evaluating rail net capacities. Technical report, RAND Corporation, Santa Monica, CA, 1955.
  • [Hu74] Te C. Hu. Optimum communication spanning trees. SIAM J. Comput., 3(3):188–195, 1974.
  • [Jan17] Svante Janson. Tail bounds for sums of geometric and exponential variables. Stat. Probab. Lett., 135, 09 2017.
  • [JKR+24] Palak Jain, Iden Kalemaj, Sofya Raskhodnikova, Satchit Sivakumar, and Adam Smith. Counting distinct elements in the turnstile model with differential privacy under continual observation. Proceedings of the Advances in Neural Information Processing Systems 36, 36, 2024.
  • [Kar93] David R. Karger. Global min-cuts in RNC, and other ramifications of a simple min-cut algorithm. In Proceedings of the 4th Annual ACM/SIGACT-SIAM Symposium on Discrete Algorithms (SODA), pages 21–30, 1993.
  • [KNRS13] Shiva Prasad Kasiviswanathan, Kobbi Nissim, Sofya Raskhodnikova, and Adam D. Smith. Analyzing graphs with node differential privacy. In Proceedings of the 10th Theory of Cryptography Conference (TCC), volume 7785, pages 457–476, 2013.
  • [Kor10] Aleksandra Korolova. Privacy violations using microtargeted ads: A case study. In Proceedings of the 2010 IEEE International Conference on Data Mining Workshops (ICDMW), pages 474–482, 2010.
  • [KRSY11] Vishesh Karwa, Sofya Raskhodnikova, Adam D. Smith, and Grigory Yaroslavtsev. Private analysis of graph structure. Proc. VLDB Endow., 4(11):1146–1157, 2011.
  • [Li21] Jason Li. Preconditioning and Locality in Algorithm Design. PhD thesis, Carnegie Mellon University, USA, 2021.
  • [LLT11] Xinyue Liu, Hongfei Lin, and Ye Tian. Segmenting webpage with Gomory-Hu tree based clustering. J. Softw., 6(12):2421–2425, 2011.
  • [LP20] Jason Li and Debmalya Panigrahi. Deterministic min-cut in poly-logarithmic max-flows. In Proceedings of the 61st IEEE Annual Symposium on Foundations of Computer Science (FOCS), pages 85–92, 2020.
  • [LP21] Jason Li and Debmalya Panigrahi. Approximate Gomory-Hu tree is faster than n1n-1 max-flows. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing (STOC), pages 1738–1748, 2021.
  • [LPS22] Jason Li, Debmalya Panigrahi, and Thatchaphol Saranurak. A nearly optimal all-pairs min-cuts algorithm in simple graphs. In Proceedings of the 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), pages 1124–1134, 2022.
  • [LUZ24a] Jingcheng Liu, Jalaj Upadhyay, and Zongrui Zou. Almost linear time differentially private release of synthetic graphs. arXiv preprint arXiv:2406.02156, 2024.
  • [LUZ24b] Jingcheng Liu, Jalaj Upadhyay, and Zongrui Zou. Optimal bounds on private graph approximation. In Proceedings of the 2024 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 1019–1049, 2024.
  • [Men27] Karl Menger. Zur allgemeinen kurventheorie. Fundam. Math., 10(1):96–115, 1927.
  • [MN21] Sagnik Mukhopadhyay and Danupon Nanongkai. A note on isolating cut lemma for submodular function minimization. arXiv preprint arXiv:2103.15724, 2021.
  • [MT07] Frank McSherry and Kunal Talwar. Mechanism design via differential privacy. In Proceedings of the 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 94–103, 2007.
  • [NS08] Arvind Narayanan and Vitaly Shmatikov. Robust de-anonymization of large sparse datasets. In Proceedings of the 2008 IEEE Symposium on Security and Privacy (SP), pages 111–125, 2008.
  • [NV21] Dung Nguyen and Anil Vullikanti. Differentially private densest subgraph detection. In Proceedings of the 38th International Conference on Machine Learning (ICML), volume 139, pages 8140–8151, 2021.
  • [Pan08] Debmalya Panigrahi. Gomory–Hu Trees, pages 364–366. Springer US, Boston, MA, 2008.
  • [PR82] Manfred W. Padberg and M. Ram Rao. Odd minimum cut-sets and b-matchings. Math. Oper. Res., 7(1):67–80, 1982.
  • [PSY22] Seth Pettie, Thatchaphol Saranurak, and Longhui Yin. Optimal vertex connectivity oracles. In Proceedings of the 54th Annual ACM SIGACT Symposium on Theory of Computing (STOC), pages 151–161, 2022.
  • [RHMS09] Vibhor Rastogi, Michael Hay, Gerome Miklau, and Dan Suciu. Relationship privacy: output perturbation for queries with joins. In Proceedings of the 28th ACM SIGMOD-SIGACT-SIGART Symposium on Principles of Database Systems (PODS), pages 107–116, 2009.
  • [RSSS21] Sofya Raskhodnikova, Satchit Sivakumar, Adam D. Smith, and Marika Swanberg. Differentially private sampling from distributions. In Proceedings of the Advances in Neural Information Processing Systems 34 (NeurIPS), pages 28983–28994, 2021.
  • [Sea16] Adam Sealfon. Shortest paths and distances with differential privacy. In Proceedings of the 35th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems (PODS), pages 29–41, 2016.
  • [SSSS17] Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership inference attacks against machine learning models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), pages 3–18, 2017.
  • [SV95] Huzur Saran and Vijay V. Vazirani. Finding kk cuts within twice the optimal. SIAM J. Comput., 24(1):101–108, 1995.
  • [US19] Jonathan R. Ullman and Adam Sealfon. Efficiently estimating erdos-renyi graphs with node differential privacy. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS), pages 3765–3775, 2019.
  • [WL93] Zhenyu Wu and Richard Leahy. An optimal graph theoretic approach to data clustering: Theory and its application to image segmentation. IEEE Trans. Pattern Anal. Mach. Intell., 15(11):1101–1113, 1993.
  • [WRRW23] Justin Whitehouse, Aaditya Ramdas, Ryan Rogers, and Steven Wu. Fully-adaptive composition in differential privacy. In Proceedings of the International Conference on Machine Learning (ICML), volume 202, pages 36990–37007, 2023.
  • [Zha21] Tianyi Zhang. Gomory-Hu trees in quadratic time. arXiv preprint arXiv:2112.01042, 2021.

Appendix A Proof of Corollary 1.3

We recall the statement.

See 1.3

Proof.

We follow the proof of [SV95]555more precisely the lecture notes in https://courses.grainger.illinois.edu/cs598csc/sp2009/lectures/lecture_7.pdf, but replace the ‘looking at’ the exact GH-tree with our approximate version. The algorithm is simple: we cut the edges corresponding to the union of cuts given by the smallest k1k-1 edges of our approximate GH-tree TT of Theorem 1.1. If this produces more than kk pieces, arbitrarily add back cut edges until we reach a kk-cut.

For the analysis, consider the optimal kk-cut with partitions V1,,VkV_{1},\ldots,V_{k} and let w(V1)w(Vk)w(V_{1})\leq\ldots\leq w(V_{k}) denote the weight of the edges leaving each partition without loss of generality. Since every edge in the optimum is adjacent to exactly two pieces of the partition, it follows that iw(Vi)\sum_{i}w(V_{i}) is twice the cost of the optimal kk-cut. We will now demonstrate k1k-1 different edges in TT which have cost at most iw(Vi)\sum_{i}w(V_{i}), up to additive error O(kΔ)=O~(nk/ε)O(k\Delta)=\tilde{O}(nk/\varepsilon), where Δ=O~(n/ε)\Delta=\tilde{O}(n/\varepsilon) is the additive error from Theorem 1.1.

As in the proof in [SV95], contract the vertices in ViV_{i} in TT for all ii. This may create parallel edges, but the resulting graph is connected since TT was connected to begin with. Make this graph into a spanning tree by removing parallel edges arbitrarily, root this graph at VkV_{k}, and orient all edges towards VkV_{k}.

Consider an arbitrary ViV_{i} where iki\neq k. The ‘supernode’ for ViV_{i} has a unique edge leaving it, which corresponds to a cut between some vertex vViv\in V_{i} with wViw\not\in V_{i}. Since TT is an approximate-GH tree, the weight of this edge must be upper bounded by w(Vi)w(V_{i}) (which is also a valid cut separating vv and ww), up to additive error Δ\Delta. The proof now follows by summing across ViV_{i}. \Box