This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Rawlsian Fairness in Online Bipartite Matching: Two-Sided, Group, and Individual

Seyed Esmaeili 1, Sharmila Duppala 1, Davidson Cheng 2, Vedant Nanda 1, Aravind Srinivasan 1, John P. Dickerson 1
Abstract

Online bipartite-matching platforms are ubiquitous and find applications in important areas such as crowdsourcing and ridesharing. In the most general form, the platform consists of three entities: two sides to be matched and a platform operator that decides the matching. The design of algorithms for such platforms has traditionally focused on the operator’s (expected) profit. Since fairness has become an important consideration that was ignored in the existing algorithms a collection of online matching algorithms have been developed that give a fair treatment guarantee for one side of the market at the expense of a drop in the operator’s profit. In this paper, we generalize the existing work to offer fair treatment guarantees to both sides of the market simultaneously, at a calculated worst case drop to operator profit. We consider group and individual Rawlsian fairness criteria. Moreover, our algorithms have theoretical guarantees and have adjustable parameters that can be tuned as desired to balance the trade-off between the utilities of the three sides. We also derive hardness results that give clear upper bounds over the performance of any algorithm.

1 Introduction

Online bipartite matching has been used to model many important applications such as crowdsourcing (Ho and Vaughan 2012; Tong et al. 2016; Dickerson et al. 2019b), rideshare (Lowalekar, Varakantham, and Jaillet 2018; Dickerson et al. 2021; Ma, Xu, and Xu 2021), and online ad allocation (Goel and Mehta 2008; Mehta 2013). In the most general version of the problem, there are three interacting entities: two sides of the market to be matched and a platform operator which assigns the matches. For example, in rideshare, riders on one side of the market submit requests, drivers on the other side of the market can take requests, and a platform operator such as Uber or Lyft matches the riders’ requests to one or more available drivers. In the case of crowdsourcing, organizations offer tasks, workers look for tasks to complete, and a platform operator such as Amazon Mechanical Turk (MTurk) or Upwork matches tasks to workers.

Online bipartite matching algorithms are often designed to optimize a performance measure—usually, maximizing overall profit for the platform operator or a proxy of that objective. However, fairness considerations were largely ignored. This is troubling especially given that recent reports have indicated that different demographic groups may not receive similar treatment. For example, in rideshare platforms once the platform assigns a driver to a rider’s request, both the rider and the driver have the option of rejecting the assignment and it has been observed that membership in a demographic group may cause adverse treatment in the form of higher rejection. Indeed, (Cook 2018; White 2016; Wirtschafter 2019) report that drivers could reject riders based on attributes such as gender, race, or disability. Conversely, (Rosenblat et al. 2016) reports that drivers are likely to receive less favorable ratings if they belong to certain demographic groups. A similar phenomenon exists in crowdsourcing (Galperin and Greppi 2017). Moreover, even in the absence of such evidence of discrimination, as algorithms become more prevalent in making decisions that directly affect the welfare of individuals (Barocas, Hardt, and Narayanan 2019; Dwork et al. 2012), it becomes important to guarantee a standard of fairness. Also, while much of our discussion focuses on the for-profit setting for concreteness, similar fairness issues hold in not-for-profit scenarios such as the fair matching of individuals with health-care facilities, e.g., in the time of a pandemic.

In response, a recent line of research has been concerned with the issue of designing fair algorithms for online bipartite matching. (Lesmana, Zhang, and Bei 2019; Ma and Xu 2022; Xu and Xu 2020) present algorithms which give a minimum utility guarantee for the drivers at a bounded drop to the operator’s profit. Conversely, (Nanda et al. 2020) give guarantees for both the platform operator and the riders instead. Finally, (Sühr et al. 2019) shows empirical methods that achieve fairness for both the riders and drivers simultaneously but lacks theoretical guarantees and ignores the operator’s profit.

Nevertheless, the existing work has a major drawback in terms of optimality guarantees. Specifically, the two sides being matched along with the platform operator constitute the three main interacting entities in online matching and despite the significant progress in fair online matching none of the previous work considers all three sides simultaneously. In this paper, we derive algorithms with theoretical guarantees for the platform operator’s profit as well as fairness guarantees for the two sides of the market. Unlike the previous work we not only consider the size of the matching but also its quality. Further, we consider two online arrival settings: the KIID and the richer KAD setting (see Section 3 for definitions). We consider both group and individual notions of Rawlsian fairness and interestingly show a reduction from individual fairness to group fairness in the KAD setting. Moreover, we show upper bounds on the optimality guarantees of any algorithm and derive impossibility results that show a conflict between group and individual notions of fairness. Finally, we empirically test our algorithms on a real-world dataset.

2 Related Work

It is worth noting that similar to our work, (Patro et al. 2020) and (Basu et al. 2020) have considered two-sided fairness as well, although in the setting of recommendation systems where a different model is applied—and, critically, a separate objective for the operator’s profit was not considered.

Fairness in bipartite matching has seen significant interest recently. The fairness definition employed has consistently been the well-known Rawlsian fairness (Rawls 1958) (i.e. max-min fairness) or its generalization Leximin fairness.***Leximin fairness maximizes the minimum utility like max-min fairness. However, it proceeds to maximize the second worst utility, and so on until the list is exhausted. We note that the objective to be maximized (other than the fairness objective) represents operator profit in our setting.

The case of offline and unweighted maximum cardinality matching is addressed by (García-Soriano and Bonchi 2020), who give an algorithm with Leximin fairness guarantees for one side of the market (one side of the bipartite graph) and show that this can be achieved without sacrificing the size of the match. Motivated by fairness consideration for drivers in ridesharing, (Lesmana, Zhang, and Bei 2019) considers the problem of offline and weighted matching. Specifically, they show an algorithm with a provable trade-off between the operator’s profit and the minimum utility guaranteed to any vertex in one-side of the market.

Recently, (Ma, Xu, and Xu 2020) considered fairness for the online part of the graph through a group notion of fairness. In particular, the utility for a group is added across the different types and is minimized for the group worst off, in rough terms their notion translates to maximizing the minimum utility accumulated by a group throughout the matching. Their notion of fairness is very similar to the one we consider here. However, (Ma, Xu, and Xu 2020) considers fairness only on one side of the graph and ignores the operator’s profit. Further, only the matching size is considered to measure utility, i.e. edges are unweighted.

A new notion of group fairness in online matching is considered in (Sankar et al. 2021). In rough terms, their group fairness criterion amounts to establishing a quota for each group and ensuring that the matching does not exceed that quota. This notion can be seen as ensuring that the system is not dominated by a specific group and is in some sense an opposite to max-min fairness as the utility is upper bounded instead of being lower bounded. Further, the fairness guarantees considered are one-sided as well.

On the empirical side of fair online matching, (Mattei, Saffidine, and Walsh 2017) and (Lee et al. 2019) give application-specific treatments in the context of deceased-donor organ allocation and food bank provisioning, respectively. More related to our work is that of (Sühr et al. 2019; Zhou, Marecek, and Shorten 2021) which consider the rideshare problem and provide algorithms to achieve fairness for both sides of the graph simultaneously, however both papers lack theoretical guarantees and in the case of (Sühr et al. 2019) the operator’s profit is not considered.

3 Online Model & Optimization Objectives

Our model follows that of (Mehta 2013; Feldman et al. 2009; Bansal et al. 2010; Alaei, Hajiaghayi, and Liaghat 2013) and others. We have a bipartite graph G=(U,V,E)G=(\mathnormal{U},\mathnormal{V},E) where UU represents the set of static (offline) vertices (workers) and VV represents the set of online vertex types (job types) which arrive dynamically in each round. The online matching is done over TT rounds. In a given round tt, a vertex of type vv is sampled from VV with probability pv,t\mathnormal{p_{v,t}} with vVpv,t=1,t[T]\sum_{v\in V}\mathnormal{p_{v,t}}=1,\forall t\in[T] the probability pv,t\mathnormal{p_{v,t}} is known beforehand for each type vv and each round tt. This arrival setting is referred to as the known adversarial distribution (KAD) setting (Alaei, Hajiaghayi, and Liaghat 2013; Dickerson et al. 2021). When the distribution is stationary, i.e. pv,t=pv,t[T]\mathnormal{p_{v,t}}=\mathnormal{p_{v}},\forall t\in[T], we have the arrival setting of the known independent identical distribution (KIID). Accordingly, the expected number of arrivals of type vv in TT rounds is nv=t[T]pv,t\mathnormal{n_{v}}=\sum_{t\in[T]}\mathnormal{p_{v,t}}, which reduces to nv=Tpv\mathnormal{n_{v}}=T\mathnormal{p_{v}} in the KIID setting. We assume that nv+\mathnormal{n_{v}}\in\mathbb{Z^{+}} for KIID (Bansal et al. 2010). Every vertex uu (vv) has a group membership,For a clearer representation we assume each vertex belongs to one group although our algorithms apply to the case where a vertex can belong to multiple groups. with 𝒢\mathcal{G} being the set of all group memberships; for any vertex uUu\in U, we denote its group memberships by g(u)𝒢g(u)\in\mathcal{G} (similarly, we have g(v)g(v) for vVv\in V). Conversely, for a group gg, U(g)U(g) (V(g)V(g)) denotes the subset of UU (VV) with group membership gg. A vertex uu (vv) has a set of incident edges EuE_{u} (EvE_{v}) which connect it to vertices in the opposite side of the graph. In a given round, once a vertex (job) vv arrives, an irrevocable decision has to be made on whether to reject vv or assign it to a neighbouring vertex uu (where (u,v)Ev(u,v)\in E_{v}) which has not been matched before. Suppose, that vv is assigned to uu, then the assignment is not necessarily successful rather it succeeds with probability pe=p(u,v)[0,1]p_{e}=p_{(u,v)}\in[0,1]. This models the fact that an assignment could fail for some reason such as the worker refusing the assigned job. Furthermore, each vertex uu has patience parameter Δu+\mathnormal{\Delta_{\mathnormal{u}}}\in\mathbb{Z^{+}} which indicates the number of failed assignments it can tolerate before leaving the system, i.e. if uu receives Δu\mathnormal{\Delta_{\mathnormal{u}}} failed assignments then it is deleted from the graph. Similarly, a vertex vv has patience Δv+\mathnormal{\Delta_{\mathnormal{v}}}\in\mathbb{Z^{+}}, if a vertex vv arrives in a given round, then it would tolerate at most Δv\mathnormal{\Delta_{\mathnormal{v}}} many failed assignments in that round before leaving the system.

For a given edge e=(u,v)Ee=(\mathnormal{u},\mathnormal{v})\in E, we let each entity assign its own utility to that edge. In particular, the platform operator assigns a utility of weO\mathnormal{w^{O}_{e}}, whereas the offline vertex u\mathnormal{u} assigns a utility of weU\mathnormal{w^{\mathnormal{U}}_{e}}, and the online vertex v\mathnormal{v} assigns a utility of weV\mathnormal{w^{\mathnormal{V}}_{e}}. This captures entities’ heterogeneous wants. For example, in ridesharing, drivers may desire long trips from nearby riders, whereas the platform operator would not be concerned with the driver’s proximity to the rider, although this maybe the only consideration the rider has. Similar motivations exist in crowdsourcing as well. We finally note that most of the details of our model such the KIID and KAD arrival settings as well as the vertex patience follow well-established and pratically motivated model choices in online matching, see Appendix (A) for more details.

Letting \mathcal{M} denote the set of successful matchings made in the TT rounds, then we consider the following optimization objectives:

  • Operator’s Utility (Profit): The operator’s expected profit is simply the expected sums of the profits across the matched edges, this leads to 𝔼[eweO]\operatorname{\mathbb{E}}[\sum_{e\in\mathcal{M}}\mathnormal{w^{O}_{e}}].

  • Rawlsian Group Fairness:

    • Offline Side: Denote by u\mathcal{M}_{\mathnormal{u}} the subset of edges in the matching that are incident on u\mathnormal{u}. Then our fairness criterion is equal to

      ming𝒢𝔼[uU(g)(euweU)]|U(g)|.\min\limits_{g\in\mathcal{G}}\frac{\operatorname{\mathbb{E}}[\sum_{u\in U(g)}(\sum_{e\in\mathcal{M}_{\mathnormal{u}}}\mathnormal{w^{\mathnormal{U}}_{e}})]}{|U(g)|}.

      this value equals the minimum average expected utility received by a group in the offline side UU.

    • Online Side: Similarly, we denote by v\mathcal{M}_{\mathnormal{v}} the subset of edges in the matching that are incident on vertex v\mathnormal{v}, and define the fairness criterion to be

      ming𝒢𝔼[vV(g)(evweV)]vV(g)nv.\min\limits_{g\in\mathcal{G}}\frac{\operatorname{\mathbb{E}}[\sum_{v\in V(g)}(\sum_{e\in\mathcal{M}_{\mathnormal{v}}}\mathnormal{w^{\mathnormal{V}}_{e}})]}{\sum_{v\in V(g)}\mathnormal{n_{v}}}.

      this value equals the minimum average expected utility received throughout the matching by any group in the online side VV.

  • Rawlsian Individual Fairness:

    • Offline Side: The definition here follows from the group fairness definition for the offline side by simply considering that each vertex uu belongs to its own distinct group. Therefore, the objective is minuU𝔼[euweU]\min\limits_{u\in U}\operatorname{\mathbb{E}}[\sum_{e\in\mathcal{M}_{\mathnormal{u}}}\mathnormal{w^{\mathnormal{U}}_{e}}].

    • Online Side: Unlike the offline side, the definition does not follow as straightforwardly. Here we cannot obtain a valid definition by simply assigning each vertex type its own group. Rather, we note that a given individual is actually a given arriving vertex at a given round t[T]t\in[T], accordingly our fairness criterion is the minimum expected utility an individual receives in a given round, i.e. mint[T]𝔼[evtweV)]\min\limits_{t\in[T]}\operatorname{\mathbb{E}}[\sum_{e\in\mathcal{M}_{\mathnormal{v}_{t}}}\mathnormal{w^{\mathnormal{V}}_{e}})], where vt\mathnormal{v}_{t} is the vertex that arrived in round tt.

4 Main Results

Performance Criterion:

We note that we are in the online setting, therefore our performance criterion is the competitive ratio. Denote by \mathcal{I} the distribution for the instances of matching problems, then OPT()=𝔼I[OPT(I)]\operatorname{\mathrm{OPT}}(\mathcal{I})=\operatorname{\mathbb{E}}_{I\sim\mathcal{I}}[\operatorname{\mathrm{OPT}}(I)] where OPT(I)\operatorname{\mathrm{OPT}}(I) is the optimal value of the sampled instance II. Similarly, for a given algorithm ALG\operatorname{\mathrm{ALG}}, we define the value of its objective over the distribution \mathcal{I} by ALG()=𝔼𝒟[ALG(I)]\operatorname{\mathrm{ALG}}(\mathcal{I})=\operatorname{\mathbb{E}}_{\mathcal{D}}[\operatorname{\mathrm{ALG}}(I)] where the expectation 𝔼𝒟[.]\operatorname{\mathbb{E}}_{\mathcal{D}}[.] is over the randomness of the instance and the algorithm. The competitive ratio is then defined as minALG()OPT()\min_{\mathcal{I}}\frac{\operatorname{\mathrm{ALG}}(\mathcal{I})}{\operatorname{\mathrm{OPT}}(\mathcal{I})}.

In our work, we address optimality guarantees for each of the three sides of the matching market by providing algorithms with competitive ratio guarantees for the operator’s profit and the fairness objectives of the static and online side of the market simultaneously. Specifically, for the KIID arrival setting we have:

Theorem 4.1.

For the KIID setting, algorithm TSGFKIID(α,β,γ)\operatorname{\mathrm{TSGF}_{\textbf{KIID}}}(\alpha,\beta,\gamma) achieves a competitive ratio of (α2e,β2e,γ2e)(\frac{\alpha}{2e},\frac{\beta}{2e},\frac{\gamma}{2e})222Here, ee denotes the Euler number, not an edge in the graph. simultaneously over the operator’s profit, the group fairness objective for the offline side, and the group fairness objective for the online side, where α,β,γ>0\alpha,\beta,\gamma>0 and α+β+γ1\alpha+\beta+\gamma\leq 1.

The following two theorems hold under the condition that pe=1,eEp_{e}=1,\forall e\in E. Specifically for the KAD setting we have:

Theorem 4.2.

For the KAD setting, algorithm TSGFKAD(α,β,γ)\operatorname{\mathrm{TSGF}_{\textbf{KAD}}}(\alpha,\beta,\gamma) achieves a competitive ratio of (α2,β2,γ2)(\frac{\alpha}{2},\frac{\beta}{2},\frac{\gamma}{2}) simultaneously over the operator’s profit, the group fairness objective for the offline side, and the group fairness objective for the online side, where α,β,γ>0\alpha,\beta,\gamma>0 and α+β+γ1\alpha+\beta+\gamma\leq 1.

Moreover, for the case of individual fairness whether in the KIID or KAD arrival setting we have:

Theorem 4.3.

For the KIID or KAD setting, we can achieve a competitive ratio of (α2,β2,γ2)(\frac{\alpha}{2},\frac{\beta}{2},\frac{\gamma}{2}) simultaneously over the operator’s profit, the individual fairness objective for the offline side, and the individual fairness objective for the online side, where α,β,γ>0\alpha,\beta,\gamma>0 and α+β+γ1\alpha+\beta+\gamma\leq 1.

We also give the following hardness results. In particular, for a given arrival (KIID or KAD) setting and fairness criterion (group or individual), the competitive ratios for all sides cannot exceed 1 simultaneously:

Theorem 4.4.

For all arrival models, given the three objectives: operator’s profit, offline side group (individual) fairness, and online side group (individual) fairness. No algorithm can achieve a competitive ratio of (α,β,γ)(\alpha,\beta,\gamma) over the three objectives simultaneously such that α+β+γ>1\alpha+\beta+\gamma>1.

It is natural to wonder if we can combine individual and group fairness. Though it is possible to extend our algorithms to this setting. The follow theorem shows that they can conflict with one another:

Theorem 4.5.

Ignoring the operator’s profit and focusing either on the offline side alone or the online side alone. With αG\mathnormal{\alpha_{G}} and αI\mathnormal{\alpha_{I}} denoting the group and individual fairness competitive ratios, respectively. No algorithm can achieve competitive ratios (αG,αI)(\mathnormal{\alpha_{G}},\mathnormal{\alpha_{I}}) over the group and individual fairness objectives of one side simultaneously such that αG+αI>1\mathnormal{\alpha_{G}}+\mathnormal{\alpha_{I}}>1.

Finally, we carry experiments on real-world datasets in Section 6.

5 Algorithms and Theoretical Guarantees

Our algorithms use linear programming (LP) based techniques (Bansal et al. 2010; Nanda et al. 2020; Xu and Xu 2020; Brubach et al. 2016b) where first a benchmark LP is set up to upper bound the optimal value of the problem, then an LP solution is sampled from to produce an algorithm with guarantees. Due to space constraints, all proofs and the technical details of Theorems (4.4 and 4.5) are in Appendix (B).

5.1 Group Fairness for the KIID Setting:

Before we discuss the details of the algorithm, we note that for a given vertex type vVv\in V, the expected arrival rate nvn_{v} could be greater than one. However, it is not difficult to modify the instance by “fragmenting” each type with nv>1n_{v}>1 such that in the new instance nv=1n_{v}=1 for each type. This can be done with the operator’s profit, offline group fairness, and online group fairness having the same values. Therefore, in what remains for the KIID setting nv=1,vVn_{v}=1,\forall v\in V and therefore for any round tt, each vertex vv arrives with probability 1T\frac{1}{T}. It also follows that for a given group gg, vV(g)nv=vV(g)1=|V(g)|\sum_{v\in V(g)}n_{v}=\sum_{v\in V(g)}1=|V(g)|.

For each edge e=(u,v)Ee=(u,v)\in E we use xex_{e} to denote the expected number of probes (i.e, assignments from uu to type vv not necessarily successful) made to edge ee in the LP benchmark. We have a total of three LPs each having the same set of constraints of (4), but differing by the objective. For compactness we do not repeat these constraints and instead write them once. Specifically, LP objective (1) along with the constraints of (4) give the optimal benchmark value of the operator’s profit. Similarly, with the same set of constraints (4) LP objective (2) and LP objective (3) give the optimal group max-min fair assignment for the offline and online sides, respectively. Note that the expected max-min objectives of (2) and (3), can be written in the form of a linear objective. For example, the max-min objective of (2) can be replaced with an LP with objective maxη\max{\eta} subject to the additional constraints that g𝒢\forall g\in\mathcal{G} , ηuU(g)eEuweUxepe|U(g)|\eta\leq\frac{\sum_{u\in U(g)}\sum_{e\in E_{u}}\mathnormal{w^{\mathnormal{U}}_{e}}x_{e}p_{e}}{|U(g)|}. Having introduced the LPs, we will use LP(1), LP(2), and LP(3) to refer to the platform’s profit LP, the offline side group fairness LP, and the online side group fairness LP, respectively.

maxeEweOxepe\displaystyle\textstyle\max\sum_{e\in E}{\mathnormal{w^{O}_{e}}x_{e}}p_{e} (1)
maxming𝒢uU(g)eEuweUxepe|U(g)|\displaystyle\textstyle\max\min\limits_{g\in\mathcal{G}}\frac{\sum_{u\in U(g)}\sum_{e\in E_{u}}\mathnormal{w^{\mathnormal{U}}_{e}}x_{e}p_{e}}{|U(g)|} (2)
maxming𝒢vV(g)eEvweVxepe|V(g)|\displaystyle\textstyle\max\min\limits_{g\in\mathcal{G}}\frac{\sum_{v\in V(g)}\sum_{e\in E_{v}}\mathnormal{w^{\mathnormal{V}}_{e}}x_{e}p_{e}}{|V(g)|} (3)
s.teE:0xe1\displaystyle\text{s.t}\quad\forall e\in E:0\leq x_{e}\leq 1 (4a)
uU:eEuxepe1\displaystyle\textstyle\forall u\in U:\sum_{e\in E_{u}}x_{e}p_{e}\leq 1 (4b)
uU:eEuxeΔu\displaystyle\textstyle\forall u\in U:\sum_{e\in E_{u}}x_{e}\leq\Delta_{u} (4c)
vV:eEvxepe1\displaystyle\textstyle\forall v\in V:\sum_{e\in E_{v}}x_{e}p_{e}\leq 1 (4d)
vV:eEvxeΔv\displaystyle\textstyle\forall v\in V:\sum_{e\in E_{v}}x_{e}\leq\Delta_{v} (4e)

Now we prove that LP(1), LP(2) and LP(3) indeed provide valid upper bounds (benchmarks) for the optimal solution for the operator’s profit and expected max-min fairness for the offline and online sides of the matching.

Lemma 5.1.

For the KIID setting, the optimal solutions of LP (1), LP (2), and LP (3) are upper bounds on the expected optimal that can be achieved by any algorithm for the operator’s profit, the offline side group fairness objective, and the online side group fairness objective, respectively.

Our algorithm makes use of the dependent rounding subroutine (Gandhi et al. 2006). We mention the main properties of dependent rounding. In particular, given a fractional vector x=(x1,x2,,xt)\vec{x}=(x_{1},x_{2},\dots,x_{t}) where each xi[0,1]x_{i}\in[0,1], let k=i[t]xik=\sum_{i\in[t]}x_{i} , dependent rounding rounds xix_{i} (possibly fractional) to Xi{0,1}X_{i}\in\{0,1\} for each i[t]i\in[t] such that the resulting vector X=(X1,X2,X3,,Xt)\vec{X}=(X_{1},X_{2},X_{3},\dots,X_{t}) has the following properties: (1) Marginal Distribution: The probability that Xi=1X_{i}=1 is equal to xix_{i}, i.e. Pr[Xi=1]=xiPr[X_{i}=1]=x_{i} for each i[t]i\in[t]. (2) Degree Preservation: Sum of XiX_{i}’s should be equal to either k\left\lfloor k\right\rfloor or k\left\lceil k\right\rceil with probability one, i.e. Pr[i[t]Xi{k,k}]=1Pr[\sum_{i\in[t]}X_{i}\in\{\left\lfloor k\right\rfloor,\left\lceil k\right\rceil\}]=1. (3) Negative Correlation: For any S[t]S\subseteq[t], (1) Pr[iSXi=0]ΠiSPr[Xi=0]Pr[\land_{i\in S}X_{i}=0]\leq\Pi_{i\in S}Pr[X_{i}=0] (2) Pr[iSXi=1]ΠiSPr[Xi=1]Pr[\land_{i\in S}X_{i}=1]\leq\Pi_{i\in S}Pr[X_{i}=1]. It follows that for any xi,xjxx_{i},x_{j}\in\vec{x}, 𝔼[Xi=1|Xj=1]xi\mathbb{E}[X_{i}=1|X_{j}=1]\leq x_{i}.

Going back to the LPs (1,2,3), we denote the optimal solutions to LP (1), LP (2), and LP (3) by x\vec{x}^{*},y\vec{y}^{*} and z\vec{z}^{*} respectively. Further, we introduce the parameters α,β,γ[0,1]\alpha,\beta,\gamma\in[0,1] where α+β+γ1\alpha+\beta+\gamma\leq 1 and each of these parameters decide the ”weight” the algorithm places on each objective (the operator’s profit, the offline group fairness, and the online group fairness objectives). We note that our algorithm makes use of the subroutine PPDR (Probe with Permuted Dependent Rounding)  shown in Algorithm 1.

Algorithm 1 PPDR(xv\vec{x}_{v})
1:  Apply dependent rounding to the fractional solution xv\vec{x}_{v} to get a binary vector Xv\vec{X}_{v}.
2:  Choose a random permutation π\pi over the set EvE_{v}.
3:  for i=1i=1 to |Ev||E_{v}| do
4:     Probe vertex π(i)\pi(i) if it is available and Xv(π(i))=1\vec{X}_{v}(\pi(i))=1
5:     if Probe is successful (i.e., a match) then
6:        break

The procedure of our parameterized sampling algorithm TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} is shown in Algorithm 2. Specifically, when a vertex of type vv arrives at any time step we run PPDR(xv\vec{x}_{v}^{*}), PPDR(yv\vec{y}_{v}^{*}), or PPDR(zv\vec{z}_{v}^{*}) with probabilities α\alpha, β\beta, and γ\gamma, respectively. We do not run any of the PPDR subroutines and instead reject the vertex with probability 1(α+β+γ)1-(\alpha+\beta+\gamma). The LP constraint (4e) guarantees that vV:eErseΔv\forall v\in V:\sum_{e\in E_{r}}s^{*}_{e}\leq\Delta_{v} where s\vec{s}^{*} could be x,y,or z\vec{x}^{*},\vec{y}^{*},\text{or }\vec{z}^{*}. Therefore, when PPDR is invoked by the degree preservation property of dependent rounding the number of edges probed will not exceed Δv\Delta_{v}, i.e. it would be within the patience limit.

Algorithm 2 TSGFKIID(α,β,γ\operatorname{\mathrm{TSGF}_{\textbf{KIID}}}(\alpha,\beta,\gamma)
1:  Let vv be the vertex type arriving at time tt.
2:  With probability α\alpha run the subroutine, PPDR(xv\vec{x}_{v}^{*}).
3:  With probability β\beta run the subroutine, PPDR(yv\vec{y}_{v}^{*}).
4:  With probability γ\gamma run the subroutine, PPDR(zv\vec{z}_{v}^{*}).
5:  Reject the arriving vertex with probability 1(α+β+γ)1-(\alpha+\beta+\gamma).

Now we analyze TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} to prove Theorem 4.1. It would suffice to prove that for each edge ee the expected number of successful probes is at least αxe2e\alpha\frac{x^{*}_{e}}{2e}, βye2e\beta\frac{y^{*}_{e}}{2e} and γze2e\gamma\frac{z^{*}_{e}}{2e}. And finally from the linearity of expectation we show that the worst case competitive ratio of the proposed online algorithm with parameters α,β\alpha,\beta and γ\gamma is at least (α2e,β2e,γ2e)(\frac{\alpha}{2e},\frac{\beta}{2e},\frac{\gamma}{2e}) for the operator’s profit and group fairness objectives on the offline and online sides of the matching, respectively.

A critical step is to lower bound the probability that a vertex uu is available (safe) at the beginning of round t[T]t\in[T]. Let us denote the indicator random variable for that event by SFu,tSF_{u,t}. The following lemma enables us to lower bound for the probability of SFu,tSF_{u,t}.

Lemma 5.2.

Pr[SFu,t](1t1T)(11T)t1Pr[SF_{u,t}]\geq\Big{(}1-\frac{t-1}{T}\Big{)}\Big{(}1-\frac{1}{T}\Big{)}^{t-1}.

Now that we have established a lower bound on Pr[SFu,t]Pr[SF_{u,t}], we lower bound the probability that an edge ee is probed by one of the PPDR subroutines conditioned on the fact that uu is available (Lemma 5.3). Let 1e,t1_{e,t} be the indicator that e=(u,v)e=(u,v) is probed by the TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} Algorithm at time tt. Note that event 1e,t1_{e,t} occurs when (1) a vertex of type vv arrives at time tt and (2) ee is sampled by PPDR(xv\vec{x_{v}}), PPDR(yv\vec{y_{v}}), or PPDR(zv\vec{z_{v}}).

Lemma 5.3.

Pr[1e,tSFu,t]αxe2TPr[1_{e,t}\mid SF_{u,t}]\geq\alpha\frac{x^{*}_{e}}{2T} ,Pr[1e,tSFu,t]βye2TPr[1_{e,t}\mid SF_{u,t}]\geq\beta\frac{y^{*}_{e}}{2T}, Pr[1e,tSFu,t]γze2TPr[1_{e,t}\mid SF_{u,t}]\geq\gamma\frac{z^{*}_{e}}{2T}

Given the above lemmas Theorem 4.1 can be proved.

5.2 Group Fairness for the KAD Setting:

For the KAD setting, the distribution over VV is time dependent and hence the probability of sampling a type vv in round tt is pv,t[0,1]\mathnormal{p_{v,t}}\in[0,1] with vVpv,t=1\sum_{v\in V}\mathnormal{p_{v,t}}=1. Further, we assume for the KAD setting that for every edge eEe\in E we have pe=1p_{e}=1. This means that whenever an incoming vertex vv is assigned to a safe-to-add vertex uu the assignment is successful. This also means that any non-trivial values for the patience parameters Δu\mathnormal{\Delta_{\mathnormal{u}}} and Δv\mathnormal{\Delta_{\mathnormal{v}}} become meaningless and hence we can WLOG assume that uU,vV,Δu=Δv=1\forall u\in U,\forall v\in V,\mathnormal{\Delta_{\mathnormal{u}}}=\mathnormal{\Delta_{\mathnormal{v}}}=1. From the above discussion, we have the following LP benchmarks for the operator’s profit, the group fairness for the offline side and the group fairness for the online side:

maxt[T]eEweOxe,t\displaystyle\textstyle\max\sum\limits_{t\in[T]}\sum\limits_{e\in E}{\mathnormal{w^{O}_{e}}\mathnormal{x_{e,t}}} (5)
maxming𝒢t[T]uU(g)eEuweUxe,t|U(g)|\displaystyle\textstyle\max\min\limits_{g\in\mathcal{G}}\frac{\sum\limits_{t\in[T]}\sum\limits_{u\in U(g)}\sum\limits_{e\in E_{u}}\mathnormal{w^{\mathnormal{U}}_{e}}\mathnormal{x_{e,t}}}{|U(g)|} (6)
maxming𝒢t[T]vV(g)eEvweVxe,tvV(g)nv\displaystyle\textstyle\max\min\limits_{g\in\mathcal{G}}\frac{\sum\limits_{t\in[T]}\sum\limits_{v\in V(g)}\sum\limits_{e\in E_{v}}\mathnormal{w^{\mathnormal{V}}_{e}}\mathnormal{x_{e,t}}}{\sum\limits_{v\in V(g)}n_{v}} (7)
s.teE,t[T]:0xe,t1\displaystyle\text{s.t}\quad\forall e\in E,\forall t\in[T]:0\leq\mathnormal{x_{e,t}}\leq 1 (8a)
uU:t[T]eEuxe,t1\displaystyle\textstyle\forall u\in U:\sum\limits_{t\in[T]}\sum\limits_{e\in E_{u}}\mathnormal{x_{e,t}}\leq 1 (8b)
vV,t[T]:eEvxe,tpv,t\displaystyle\textstyle\forall v\in V,\forall t\in[T]:\sum_{e\in E_{v}}\mathnormal{x_{e,t}}\leq\mathnormal{p_{v,t}} (8c)
Lemma 5.4.

For the KAD setting, the optimal solutions of LP (5), LP (6) and LP (7) are upper bounds on the expected optimal that can be achieved by any algorithm for the operator’s profit, the offline side group fairness objective, and the online side group fairness objective, respectively.

Note that in the above LP we have xe,t\mathnormal{x_{e,t}} as the probability for successfully assigning an edge in round tt (with an explicit dependence on tt), unlike in the KIID setting where we had xex_{e} instead to denote the expected number of times edge ee is probed through all rounds. Similar to our solution for the KIID setting, we denote by xe,t\mathnormal{x^{*}_{e,t}}, ye,t\mathnormal{y^{*}_{e,t}}, and ze,t\mathnormal{z^{*}_{e,t}} the optimal solutions of the LP benchmarks for the operator’s profit, offline side group fairness, and online side group fairness, respectively.

Having the optimal solutions to the LPs, we use algorithm TSGFKAD\operatorname{\mathrm{TSGF}_{\textbf{KAD}}} shown in Algorithm 3. In TSGFKAD\operatorname{\mathrm{TSGF}_{\textbf{KAD}}} new parameters are introduced, specifically λ\lambda and ρe,t\mathnormal{\rho_{e,t}} where ρe,t\mathnormal{\rho_{e,t}} is the probability that edge e=(u,v)e=(u,v) is safe to add in round tt, i.e. the probability that uu is unmatched at the beginning of round tt. For now we assume that we have the precise values of ρe,t\mathnormal{\rho_{e,t}} for all rounds and discuss how to obtain these values at the end of this subsection. Now conditioned on vv arriving at round tt and e=(u,v)e=(u,v) being safe to add, it follows that ee is sampled with probability αxe,tpv,tλρe,t+βye,tpv,tλρe,t+γze,tpv,tλρe,t\alpha\frac{\mathnormal{x^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\beta\frac{\mathnormal{y^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\gamma\frac{\mathnormal{z^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}} which would be a valid probability (positive and not exceeding 1) if ρe,tλ\mathnormal{\rho_{e,t}}\geq\lambda. This follows from the fact that α,β,γ[0,1]\alpha,\beta,\gamma\in[0,1] and α+β+γ1\alpha+\beta+\gamma\leq 1 and also by constraint (8c) which leads to eEvxe,tpv,t1\frac{\sum_{e\in E_{v}}\mathnormal{x_{e,t}}}{\mathnormal{p_{v,t}}}\leq 1. Further, if ρe,tλ\mathnormal{\rho_{e,t}}\geq\lambda then by constraint (8c) we have eEv(αxe,tpv,tλρe,t+βye,tpv,tλρe,t+γze,tpv,tλρe,t)1\sum_{e\in E_{v}}\Big{(}\alpha\frac{\mathnormal{x^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\beta\frac{\mathnormal{y^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\gamma\frac{\mathnormal{z^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}\Big{)}\leq 1 and therefore the distribution is valid. Clearly, the value of λ\lambda is important for the validity of the algorithm, the following lemma shows that λ=12\lambda=\frac{1}{2} leads to a valid algorithm.

Lemma 5.5.

Algorithm TSGFKAD\operatorname{\mathrm{TSGF}_{\textbf{KAD}}} is valid for λ=12\lambda=\frac{1}{2}.

Algorithm 3 TSGFKAD(α,β,γ\operatorname{\mathrm{TSGF}_{\textbf{KAD}}}(\alpha,\beta,\gamma)
1:  Let vv be the vertex type arriving at time tt.
2:  if  Ev,t=ϕE_{v,t}=\phi then
3:     Reject vv
4:  else
5:     With probability α\alpha probe ee with probability xe,tpv,tλρe,t\frac{\mathnormal{x^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}.
6:     With probability β\beta probe ee with probability ye,tpv,tλρe,t\frac{\mathnormal{y^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}.
7:     With probability γ\gamma probe ee with probability ze,tpv,tλρe,t\frac{\mathnormal{z^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}.
8:     With probability [1(α+β+γ)][1-(\alpha+\beta+\gamma)] reject vv .

We now return to the issue of how to obtain the values of ρe,t\mathnormal{\rho_{e,t}} for all rounds. This can be done by using the simulation technique as done in (Dickerson et al. 2021; Adamczyk, Grandoni, and Mukherjee 2015). To elaborate, we note that we first solve the LPs (5,6,7) and hence have the values of xe,t\mathnormal{x^{*}_{e,t}}, ye,t\mathnormal{y^{*}_{e,t}}, and ze,t\mathnormal{z^{*}_{e,t}}. Now, for the first round t=1t=1, clearly ρe,t=1,eE\mathnormal{\rho_{e,t}}=1,\forall e\in E. To obtain ρe,t\mathnormal{\rho_{e,t}} for t=2t=2, we simulate the arrivals and algorithm a collection of times, and use the empirically estimated probability. More precisely, for t=1t=1 we sample the arrival of vertex vv from pv,t\mathnormal{p_{v,t}} with t=1t=1 (pv,t\mathnormal{p_{v,t}} values are given as part of the model), then we run our algorithm for the values of α,β,γ\alpha,\beta,\gamma that we have chosen. Accordingly, at t=2t=2 some vertex in UU might be matched. We do this simulation a number of times and then we take ρe,t\mathnormal{\rho_{e,t}} for t=2t=2 to be the average of all runs. Now having the values of ρe,t\mathnormal{\rho_{e,t}} for t=1t=1 and t=2t=2, we further simulate the arrivals and the algorithm to obtain ρe,t\mathnormal{\rho_{e,t}} for t=3t=3 and so on until we get ρe,t\mathnormal{\rho_{e,t}} for the last round TT. We note that using the Chernoff bound (Mitzenmacher and Upfal 2017) we can rigorously characterize the error in this estimation, however by doing this simulation a number of times that is polynomial in the problem size, the error in the estimation would only affect the lower order terms in the competitive ration analysis (Dickerson et al. 2021) and hence for simplicity it is ignored. Now, with Lemma 5.5 Theorem 4.2 can be proved (see Appendix (B)).

5.3 Individual Fairness KIID and KAD Settings:

For the case of Rawlsian (max-min) individual fairness, we consider each vertex of the offline side to belong to its own distinct group and the definition of group max-min fairness would lead to individual max-min fairness. On the other hand, for the online side a similar trick would not yield a meaningful criterion, we instead define the individual max-min fairness for the online side to equal mint[T]𝔼[util(vt)]=mint[T]𝔼[evtweV)]\min\limits_{t\in[T]}\operatorname{\mathbb{E}}[\text{util}(v_{t})]=\min\limits_{t\in[T]}\operatorname{\mathbb{E}}[\sum_{e\in\mathcal{M}_{\mathnormal{v}_{t}}}\mathnormal{w^{\mathnormal{V}}_{e}})] where util(vt)\text{util}(v_{t}) is the utility received by the vertex arriving in round tt. If we were to denote by xe,t\mathnormal{x_{e,t}} the probability that the algorithm would successfully match ee in round tt, then it follows straightforwardly that 𝔼[util(vt)]=eEvtweVxe,t\operatorname{\mathbb{E}}[\text{util}(v_{t})]=\sum_{e\in E_{v_{t}}}\mathnormal{w^{\mathnormal{V}}_{e}}\mathnormal{x_{e,t}}. We consider this definition to be the valid extension of max-min fairness for the online side as we are now concerned with the minimum utility across the online individuals (arriving vertices) which are TT many. The following lemma shows that we can solve two-sided individual max-min fairness by a reduction to two-sided group max-min fairness in the KAD arrival setting:

Lemma 5.6.

Whether in the KIID or KAD setting, a given instance of two-sided individual max-min fairness can be converted to an instance of two-sided group max-min fairness in the KAD setting.

The above Lemma with algorithm TSGFKAD\operatorname{\mathrm{TSGF}_{\textbf{KAD}}} can be used to prove Theorem 4.3 as shown in Appendix (B).

6 Experiments

Refer to caption
Figure 1: Competitive ratios for TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} over the operator’s profit, offline (driver) fairness objective, and online (rider) fairness objective with different values of α,β,γ\alpha,\beta,\gamma. Note that “Matching” refers to the case where driver and rider utilities are set to 1 across all edges. The experiment is run with α={0,0.1,0.2,,1}\alpha=\{0,0.1,0.2,...,1\}, and β=γ=1α2\beta=\gamma=\frac{1-\alpha}{2}. Higher competitive ratio indicates better performance.

In this section, we verify the performance of our algorithm and our theoretical lower bounds for the KIID and group fairness setting using algorithm TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} (Section 5.1). We note that none of the previous work consider our three-sided setting. We use rideshare as an application example of online bipartite matching (see also, e.g., Dickerson et al. 2021; Nanda et al. 2020; Xu and Xu 2020; Barann, Beverungen, and Müller 2017). We expect similar results and performance to hold in other matching applications such as crowdsourcing.

Experimental Setup:

As done in previous work, the drivers’ side is the offline (static) side whereas the riders’ side is the online side. We run our experiments over the widely used New York City (NYC) yellow cabs dataset (Sekulić, Long, and Demšar 2021; Nanda et al. 2020; Xu and Xu 2020; Alonso-Mora, Wallar, and Rus 2017) which contains records of taxi trips in the NYC area from 2013. Each record contains a unique (anonymized) ID of the driver, the coordinates of start and end locations of the trip, distance of the trip, and additional metadata.

Similar to (Dickerson et al. 2021; Nanda et al. 2020), we bin the starting and ending latitudes and longitudes by dividing the latitudes from 40.440.4^{\circ} to 40.9540.95^{\circ} and longitudes from 73-73^{\circ} to 75-75^{\circ} into equally spaced grids of step size 0.0050.005. This enables us to define each driver and request type based on its starting and ending bins. We pick out the trips between 7pm and 8pm on January 31, 2013, which is a rush hour with 10,814 drivers and 35,109 trips. We set driver patience Δu\Delta_{u} to 3. Following (Xu and Xu 2020), we uniformly sample rider patience Δv\Delta_{v} from {1,2}\{1,2\}.

Since the dataset does not include demographic information, for each vertex we randomly sample the group membership (Nanda et al. 2020). Specifically, we randomly assign 70%70\% of the riders and drivers to be advantaged and the rest to be disadvantaged. The value of pep_{e} for e=(u,v)e=(u,v) depends on whether the vertices belong to the advantaged or disadvantaged group. Specifically, pe=0.6p_{e}=0.6 if both vertices are advantaged, pe=0.3p_{e}=0.3 if both are disadvantaged, and pe=0.1p_{e}=0.1 for other cases.

In addition to this, a key component of our work is the use of driver and rider specific utilities. We follow the work of (Sühr et al. 2019) to set the utilities. We adopt the Manhattan distance metric rather than the Euclidean distance metric since the former is a better proxy for length of taxi trips in New York City. We set the operator’s utility to the rider’s trip length weO=tripLength(v)w^{O}_{e}=\text{tripLength$(v)$}—a rough proxy for profit. In addition, the rider’s utility over an edge e=(u,v)e=(u,v) is set to weV=dist(u,v)w^{V}_{e}=-\text{dist$(u,v)$} where dist(u,v)(u,v) is the distance between the rider and the driver. The driver’s utility is set to weU=tripLength(v)dist(u,v)w^{U}_{e}=\text{tripLength$(v)$}-\text{dist$(u,v)$}. Whereas the trip length tripLength(v)(v) is available in the dataset, the distance between the rider and the driver dist(u,v)(u,v) is not. We therefore simulate the distance, by creating an equally spaced grid with step size 0.0050.005 around the starting coordinates of the trip. This results in 81 possible coordinates in the vicinity of the starting coordinates of the trip. We then randomly choose one of these 81 coordinates to be the location of the driver when the trip was requested. Then dist(u,v)dist(u,v) is the distance between this coordinate to the start coordinate of the trip. This is a valid approximation since the platform would not assign drivers unreasonably far away to pickup a rider. Lastly, we scale the utilities by a constant to prevent them from being negative.

We run TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} at the scale of |U|=49|U|=49, |V|=172|V|=172 for 100 trials. During each trial, we randomly sample 49 drivers and 172 requests between 7 and 8pm, and run TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} 100 times to measure the expected competitive ratios of this trial. We then averaged the competitive ratios over all trials, and the results are reported in figure 1. Code to reproduce our experiments is available in the blinded formathttps://github.com/anonymousUser634534/TSGF; we will release that code in deblinded form upon acceptance.

Performance of TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} with Varied Parameters:

Figure 1 shows the performance of our algorithm over the three objectives: operator’s profit, offline (driver) group fairness, and online (rider) group fairness. It is clear that the algorithm behaves as expected with all objectives being steadily above their theoretical lower bound. More importantly, we see that increasing the weight for an objective leads to better performance for that objective. I.e., a higher weight for β\beta leads to better performance for the offline side fairness and similar observations follow in the case of α\alpha for the operator’s objective and in the case of γ\gamma for the online-fairness. This also indicates the limitation in previous work which only considered fairness for one-side since their algorithms would not be able to improve the fairness for the other ignored side.

Furthermore, previous work (e.g., Nanda et al. 2020; Xu and Xu 2020; Ma and Xu 2022) only considered the matching size when optimizing the fairness objective for the offline (drivers) or online (riders) side. This is in contrast to our setting where we consider the matching quality. To see the effect of ignoring the matching quality and only considering the size, we run the same experiments with weU=weV=1,eEw^{U}_{e}=w^{V}_{e}=1,\forall e\in E, i.e. the quality is ignored. The results are shown shown in the graph labelled “Matching” in figure 1, it is clear that ignoring the match quality leads to noticeably worse results.

Comparison to Heuristics:

We also compare the performance of TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} against three other heuristics. In particular, we consider Greedy-O which is a greedy algorithm that upon the arrival of an online vertex (rider) vv picks the edge eEve\in E_{v} with maximum value of peweOp_{e}w^{O}_{e} until it either results in a match or the patience quota is reached. We also consider Greedy-R which is identical to Greedy-O except that it greedily picks the edge with maximum value of peweVp_{e}w^{V}_{e} instead, therefore maximizing the rider’s utility in a greedy fashion. Moreover, we consider Greedy-D which is a greedy algorithm that upon the arrival of an online vertex vv, first finds the group on the offline side with the lowest average utility so far, then it greedily picks an offline vertex (driver) uEvu\in E_{v} from this group (if possible) which has the maximum utility until it either results in a match or the patience limit is reached. We carried out 100 trials to compare the performance of TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} with the greedy algorithms, where each trial contains 49 randomly sampled drivers and 172 requests and is repeated 100 times. The aggregated results are displayed in table 1. We see that TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} outperforms the heuristics with the exception of a small under-performance in comparison to Greedy-D. However, using Greedy-D we cannot tune the weights (α\alpha, β\beta, and γ\gamma) to balance the objectives as we can in the case of TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}}.

Profit
Driver
Fairness
Rider
Fairness
Greedy-O 0.431 0.549 0.503
TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} (α=1\alpha=1) 0.595 0.398 0.384
Greedy-D 0.371 0.609 0.563
TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} (β=1\beta=1) 0.517 0.571 0.44
Greedy-R 0.316 0.504 0.513
TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} (γ=1\gamma=1) 0.252 0.353 0.574
Table 1: Competitive ratios of TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} with Greedy heuristics on the NYC dataset at |U|=49|U|=49, |V|=172|V|=172. Higher competitive ratio indicates better performance.

Acknowledgments

This research was supported in part by NSF CAREER Award IIS-1846237, NSF Award CCF-1918749, NSF Award CCF-1852352, NSF Award SMA-2039862, NIST MSE Award #20126334, DARPA GARD #HR00112020007, DARPA SI3-CMD #S4761, DoD WHS Award #HQ003420F0035, ARPA-E DIFFERENTIATE Award #1257037, ARL Award W911NF2120076, and gifts by research awards from Amazon, and Google. We are grateful to Pan Xu for advice and comments on earlier versions of this work.

References

  • Adamczyk, Grandoni, and Mukherjee (2015) Adamczyk, M.; Grandoni, F.; and Mukherjee, J. 2015. Improved approximation algorithms for stochastic matching. In Algorithms-ESA 2015, 1–12. Springer.
  • Alaei, Hajiaghayi, and Liaghat (2012) Alaei, S.; Hajiaghayi, M.; and Liaghat, V. 2012. Online prophet-inequality matching with applications to ad allocation. In Proceedings of the 13th ACM Conference on Electronic Commerce, 18–35.
  • Alaei, Hajiaghayi, and Liaghat (2013) Alaei, S.; Hajiaghayi, M.; and Liaghat, V. 2013. The online stochastic generalized assignment problem. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, 11–25. Springer.
  • Alonso-Mora, Wallar, and Rus (2017) Alonso-Mora, J.; Wallar, A.; and Rus, D. 2017. Predictive routing for autonomous mobility-on-demand systems with ride-sharing. In 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 3583–3590.
  • Bahmani and Kapralov (2010) Bahmani, B.; and Kapralov, M. 2010. Improved bounds for online stochastic matching. In European Symposium on Algorithms, 170–181. Springer.
  • Bansal et al. (2010) Bansal, N.; Gupta, A.; Li, J.; Mestre, J.; Nagarajan, V.; and Rudra, A. 2010. When lp is the cure for your matching woes: Improved bounds for stochastic matchings. In European Symposium on Algorithms, 218–229. Springer.
  • Barann, Beverungen, and Müller (2017) Barann, B.; Beverungen, D.; and Müller, O. 2017. An open-data approach for quantifying the potential of taxi ridesharing. Decision Support Systems, 99: 86–95.
  • Barocas, Hardt, and Narayanan (2019) Barocas, S.; Hardt, M.; and Narayanan, A. 2019. Fairness and Machine Learning. fairmlbook.org. Accessed: 2022-08-01.
  • Basu et al. (2020) Basu, K.; DiCiccio, C.; Logan, H.; and Karoui, N. E. 2020. A Framework for Fairness in Two-Sided Marketplaces. arXiv preprint arXiv:2006.12756.
  • Brubach et al. (2016a) Brubach, B.; Sankararaman, K. A.; Srinivasan, A.; and Xu, P. 2016a. New algorithms, better bounds, and a novel model for online stochastic matching. In 24th Annual European Symposium on Algorithms (ESA 2016). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik.
  • Brubach et al. (2016b) Brubach, B.; Sankararaman, K. A.; Srinivasan, A.; and Xu, P. 2016b. Online Stochastic Matching: New Algorithms and Bounds. arXiv:1606.06395.
  • Cook (2018) Cook, G. 2018. Woman Says Uber Driver Denied Her Ride Because of Her Wheelchair. NBC4-Washington. Available at https://www.nbcwashington.com/news/local/woman-says-uber-driver-denied-her-ride-because-of-her-wheelchair/2029780/, Accessed: 2023-04-23.
  • Dickerson et al. (2019a) Dickerson, J.; Sankararaman, K.; Sarpatwar, K.; Srinivasan, A.; Wu, K.-L.; and Xu, P. 2019a. Online resource allocation with matching constraints. In International Conference on Autonomous Agents and Multiagent Systems (AAMAS).
  • Dickerson et al. (2019b) Dickerson, J. P.; Sankararaman, K. A.; Srinivasan, A.; and Xu, P. 2019b. Balancing relevance and diversity in online bipartite matching via submodularity. In AAAI.
  • Dickerson et al. (2021) Dickerson, J. P.; Sankararaman, K. A.; Srinivasan, A.; and Xu, P. 2021. Allocation Problems in Ride-sharing Platforms: Online Matching with Offline Reusable Resources. ACM Transactions on Economics and Computation (TEAC), 9(3): 1–17.
  • Dwork et al. (2012) Dwork, C.; Hardt, M.; Pitassi, T.; Reingold, O.; and Zemel, R. 2012. Fairness through awareness. In ITCS.
  • Feldman et al. (2009) Feldman, J.; Mehta, A.; Mirrokni, V.; and Muthukrishnan, S. 2009. Online stochastic matching: Beating 1-1/e. In 2009 50th Annual IEEE Symposium on Foundations of Computer Science, 117–126. IEEE.
  • Galperin and Greppi (2017) Galperin, H.; and Greppi, C. 2017. Geographical discrimination in the gig economy. Available at SSRN 2922874.
  • Gandhi et al. (2006) Gandhi, R.; Khuller, S.; Parthasarathy, S.; and Srinivasan, A. 2006. Dependent rounding and its applications to approximation algorithms. Journal of the ACM (JACM), 53(3): 324–360.
  • García-Soriano and Bonchi (2020) García-Soriano, D.; and Bonchi, F. 2020. Fair-by-design matching. Data Mining and Knowledge Discovery, 1–45.
  • Goel and Mehta (2008) Goel, G.; and Mehta, A. 2008. Online budgeted matching in random input models with applications to Adwords. In SODA.
  • Ho and Vaughan (2012) Ho, C.-J.; and Vaughan, J. 2012. Online task assignment in crowdsourcing markets. In AAAI.
  • Karp, Vazirani, and Vazirani (1990) Karp, R. M.; Vazirani, U. V.; and Vazirani, V. V. 1990. An optimal algorithm for on-line bipartite matching. In Proceedings of the twenty-second annual ACM symposium on Theory of computing, 352–358.
  • Lee et al. (2019) Lee, M. K.; Kusbit, D.; Kahng, A.; Kim, J. T.; Yuan, X.; Chan, A.; See, D.; Noothigattu, R.; Lee, S.; Psomas, A.; ; and Procaccia, A. D. 2019. WeBuildAI: Participatory framework for algorithmic governance. In CSCW.
  • Lesmana, Zhang, and Bei (2019) Lesmana, N. S.; Zhang, X.; and Bei, X. 2019. Balancing efficiency and fairness in on-demand ridesourcing. In NeurIPS.
  • Lowalekar, Varakantham, and Jaillet (2018) Lowalekar, M.; Varakantham, P.; and Jaillet, P. 2018. Online spatio-temporal matching in stochastic and dynamic domains. Artificial Intelligence (AIJ), 261: 71–112.
  • Ma and Xu (2022) Ma, W.; and Xu, P. 2022. Group-level fairness maximization in online bipartite matching. In AAMAS.
  • Ma, Xu, and Xu (2020) Ma, W.; Xu, P.; and Xu, Y. 2020. Group-level fairness maximization in online bipartite matching. arXiv preprint arXiv:2011.13908.
  • Ma, Xu, and Xu (2021) Ma, W.; Xu, P.; and Xu, Y. 2021. Fairness Maximization among Offline Agents in Online-Matching Markets. In WINE.
  • Manshadi, Gharan, and Saberi (2012) Manshadi, V. H.; Gharan, S. O.; and Saberi, A. 2012. Online stochastic matching: Online actions based on offline statistics. Mathematics of Operations Research, 37(4): 559–573.
  • Mattei, Saffidine, and Walsh (2017) Mattei, N.; Saffidine, A.; and Walsh, T. 2017. Mechanisms for Online Organ Matching. In IJCAI.
  • Mehta (2013) Mehta, A. 2013. Online Matching and Ad Allocation. Foundations and Trends in Theoretical Computer Science, 8(4): 265–368.
  • Mitzenmacher and Upfal (2017) Mitzenmacher, M.; and Upfal, E. 2017. Probability and computing: Randomization and probabilistic techniques in algorithms and data analysis. Cambridge university press.
  • Nanda et al. (2020) Nanda, V.; Xu, P.; Sankararaman, K. A.; Dickerson, J. P.; and Srinivasan, A. 2020. Balancing the Tradeoff between Profit and Fairness in Rideshare Platforms During High-Demand Hours. In AAAI.
  • Patro et al. (2020) Patro, G. K.; Biswas, A.; Ganguly, N.; Gummadi, K. P.; and Chakraborty, A. 2020. FairRec: Two-Sided Fairness for Personalized Recommendations in Two-Sided Platforms. In Proceedings of The Web Conference 2020, 1194–1204.
  • Rawls (1958) Rawls, J. 1958. Justice as fairness. The philosophical review, 67(2): 164–194.
  • Rosenblat et al. (2016) Rosenblat, A.; Levy, K. E.; Barocas, S.; and Hwang, T. 2016. Discriminating tastes: Customer ratings as vehicles for bias. Available at SSRN 2858946.
  • Sankar et al. (2021) Sankar, G. S.; Louis, A.; Nasre, M.; and Nimbhorkar, P. 2021. Matchings with Group Fairness Constraints: Online and Offline Algorithms. arXiv preprint arXiv:2105.09522.
  • Sekulić, Long, and Demšar (2021) Sekulić, S.; Long, J.; and Demšar, U. 2021. A spatially aware method for mapping movement-based and place-based regions from spatial flow networks. Transactions in GIS, 25(4): 2104–2124.
  • Sühr et al. (2019) Sühr, T.; Biega, A. J.; Zehlike, M.; Gummadi, K. P.; and Chakraborty, A. 2019. Two-sided fairness for repeated matchings in two-sided markets: A case study of a ride-hailing platform. In KDD.
  • Tong et al. (2016) Tong, Y.; She, J.; Ding, B.; Wang, L.; and Chen, L. 2016. Online mobile micro-task allocation in spatial crowdsourcing. In ICDE.
  • White (2016) White, G. B. 2016. Uber and Lyft Are Failing Black Riders. The Atlantic. Available at https://www.theatlantic.com/business/archive/2016/10/uber-lyft-and-the-false-promise-of-fair-rides/506000/, Accessed: 2023-04-22.
  • Wirtschafter (2019) Wirtschafter, E. 2019. Driver discrimination still a problem as Uber and Lyft prepare to go public. KALW. Available at https://www.kalw.org/post/driver-discrimination-still-problem-uber-and-lyft-prepare-go-public, Accessed: 2023-04-23.
  • Xu and Xu (2020) Xu, Y.; and Xu, P. 2020. Trade the System Efficiency for the Income Equality of Drivers in Rideshare. In IJCAI.
  • Zhou, Marecek, and Shorten (2021) Zhou, Q.; Marecek, J.; and Shorten, R. N. 2021. Subgroup Fairness in Two-Sided Markets. arXiv preprint arXiv:2106.02702.

Appendix A Online Matching Model Details

A.1 Arrival Setting (KIID and KAD):

The modelling choices we have made follow standard settings in online matching (Mehta 2013; Alaei, Hajiaghayi, and Liaghat 2013). To elaborate further, the initial seminal paper on online matching (Karp, Vazirani, and Vazirani 1990) does not assume any prior knowledge on the arrival of the online vertices of VV and follows adversarial analysis to establish theoretical guarantees on the competitive ratio. In addition to overly pessimistic theoretical results, the lack of prior knowledge is often an unrealistic assumption. Most decision makers in online matching settings are able to gain knowledge on the arrival rates of the online vertices and this knowledge can be used to build more realistic probabilistic knowledge of the arrival.

Specifically, the Known Independent and Identically Distributed KIID model is an established model in online matching (Feldman et al. 2009; Mehta 2013; Bahmani and Kapralov 2010; Manshadi, Gharan, and Saberi 2012; Dickerson et al. 2019b). In this model, the collection of arriving vertices on the online side belong to a finite set of known types where the type of a vertex vv decides the edge connections EvE_{v} it has to the vertices of UU along with the weights we,eEvw_{e},\forall e\in E_{v} of those edges. Further, a given vertex of type vv arrives with the same probability pvp_{v} in every round. These arrival probabilities can be estimated easily from historical data based on previous matchings.

While the KIID model utilizes prior knowledge which is frequently available in practical applications, it is still restrictive since it assumes that the probabilities do not vary through time. The Known Adversarial Arrival KAD model (also known as prophet inequality matching) on the other hand, takes into account the dynamic variation in the probabilities. Therefore, the probability a vertex of type vv arrives in round tt is pv,tp_{v,t} instead of being constant for every round tt. This model is also well-established in the matching literature and has been used in a collection of papers such as (Alaei, Hajiaghayi, and Liaghat 2012; Brubach et al. 2016a; Dickerson et al. 2021, 2019a). Despite the fact that the KAD model is well-motivated and richer than the KIID model it was not used in the one-sided online fair matching papers of (Nanda et al. 2020; Xu and Xu 2020).

A.2 Patience:

The patience parameter of a vertex Δu\Delta_{u} (or Δv\Delta_{v}) for an offline vertex uu (or an online vertex vv) models its tolerance for unsuccessful probes (match attempts) before leaving the system. We note that this is an important detail in the online matching model since it is frequently the case that the vertices in the online matching applications (such as advertising, crowdsourcing, and ridesharing) represent human participants who would only tolerate a fixed number of failed matching attempts before leaving the system. Like the KIID and KAD arrival models, the patience parameter is also well-established in online matching, see for example (Mehta 2013; Bansal et al. 2010; Adamczyk, Grandoni, and Mukherjee 2015). Despite the importance of this parameter, the previous work in fair online matching did not consider the patience issue for both sides simultaneously (Nanda et al. 2020; Xu and Xu 2020), handling both parameters at the same time is more challenging and leads to more tedious derivations.

We further elaborate on the meaning of the patience for both the online and offline sides, we note again that this is following the research literature on online matching:

Offline Patience:

Consider a vertex uu with patience Δu\Delta_{u}, then vertex uu will remain on the offline side UU unless it is successfully matched or it receives Δu\Delta_{u} many failed matching attempts. As a concrete example, consider a vertex u1u_{1} with patience Δu1=2\Delta_{u_{1}}=2. Clearly, in the first round (t=1t=1) u1u_{1} will be in the offline side UU, suppose an unsuccessful matching attempt (unsuccessful probe) is made in this round, then in the next round u1u_{1} will still be there. Suppose that the next round when u1u_{1} is probed is in the fifth round (t=5t=5), then if the probe is successful then u1u_{1} is matched and will be removed from the offline side in the next rounds (t>5t>5), but also if the match is unsuccessful then u1u_{1} will not be matched but will still be removed for all of the next rounds (t>5t>5) since it has a patience Δu1=2\Delta_{u_{1}}=2 and therefore can only take two failed matching attempts before leaving.

Online Patience:

Unlike the offline side, an online vertex vv would arrive in a round tt and must be matched or rejected in that given round. While in a round tt we can at most match one online vertex (which is the arriving vertex vv) to some offline vertex uu, we can make multiple match attempts (probes) from vv to the vertices it is connected to in UU in that round tt. The patience Δv\Delta_{v} of vv decides the upper limit on the number of failed attempts we can make in round tt before vv leaves the system and can no longer be matched even if a possible match was still not attempted. As a concrete example, suppose vertex of type v1v_{1} with Δv1=3\Delta_{v_{1}}=3 arrives in round t=7t=7 and that v1v_{1} is connected to a total of four vertices {u1,u2,u3,u4}\{u_{1},u_{2},u_{3},u_{4}\} in UU all of which are still available (i.e. unmatched and still have not ran out of patience), suppose we make match attempts (probes) to u1u_{1} then u2u_{2} then u3u_{3}, it follows since Δv1=3\Delta_{v_{1}}=3 that v1v_{1} has left the system and we can no longer even attempt to match it to u4u_{4} despite that fact that its available. Further, if at any probe attempt v1v_{1} was matched then no further probe attempts are made to v1v_{1}, e.g. if the first probe (v1v_{1} to u1u_{1}) in the above discussion was successful, then v1v_{1} and u1u_{1} are matched to each other and we cannot attempt to match v1v_{1} to u2,u3,or u4u_{2},u_{3},\text{or }u_{4}.

Appendix B Proofs

Here we include the missing proofs. Each lemma/theorem is restated followed by its proof.

B.1 Proofs for Section 5.1

See 5.1

Proof.

We follow a similar proof to that used in (Bansal et al. 2010). We shall focus on the operator’s profit objective since the other objectives follow by very similar arguments. First, we note that LP(1) uses the expected values of the problem parameters, i.e. if we consider a specific graph realization GG, then let NvGN^{G}_{v} be the number of arrival for vertex type vv, then it follows that LP(1) uses the expected values since 𝔼[NvG]=1,vV\operatorname{\mathbb{E}}_{\mathcal{I}}[N^{G}_{v}]=1,\forall v\in V where 𝔼[.]\operatorname{\mathbb{E}}_{\mathcal{I}}[.] is an expectation over the randomness of the instance. We shall therefore refer to the value of LP(1) as LP(𝔼[G])LP(\operatorname{\mathbb{E}_{\mathcal{I}}}[G]).

To prove that LP(𝔼(G))LP(\operatorname{\mathbb{E}_{\mathcal{I}}}(G)) is a valid upper bound it suffices to show that LP(𝔼[G[)𝔼[LP(G)]LP(\operatorname{\mathbb{E}_{\mathcal{I}}}[G[)\geq\operatorname{\mathbb{E}_{\mathcal{I}}}[LP(G)] where LP(G)LP(G) is the optimal LP value of a realized instance GG and 𝔼[LP(G)]\operatorname{\mathbb{E}_{\mathcal{I}}}[LP(G)] is the expected value of that optimal LP value. Let us then consider a specific realization GG^{\prime}, its corresponding LP would be the following:

maxeEweOpexe\displaystyle\textstyle\max\sum_{e^{\prime}\in E^{\prime}}{w^{O}_{e^{\prime}}p_{e^{\prime}}x_{e^{\prime}}} (9)
s.teE:0xe1\displaystyle\text{s.t}\quad\forall e^{\prime}\in E^{\prime}:0\leq x_{e^{\prime}}\leq 1 (10a)
uU:eEuxepe1\displaystyle\textstyle\forall u\in U:\sum_{e^{\prime}\in E^{\prime}_{u}}x_{e^{\prime}}p_{e^{\prime}}\leq 1 (10b)
uU:eEuxeΔu\displaystyle\textstyle\forall u\in U:\sum_{e^{\prime}\in E^{\prime}_{u}}x_{e^{\prime}}\leq\Delta_{u} (10c)
vV:eEvxepe1\displaystyle\textstyle\forall v^{\prime}\in V^{\prime}:\sum_{e^{\prime}\in E^{\prime}_{v^{\prime}}}x_{e^{\prime}}p_{e^{\prime}}\leq 1 (10d)
vV:eEvxeΔv\displaystyle\textstyle\forall v^{\prime}\in V^{\prime}:\sum_{e^{\prime}\in E^{\prime}_{v^{\prime}}}x_{e^{\prime}}\leq\Delta_{v^{\prime}} (10e)

where VV^{\prime} is the realization of the online side. It is clear that for a given realization G=(U,V,E)G^{\prime}=(U,V^{\prime},E^{\prime}) the above LP(9) is an upper bound on the operator’s objective value for that realization.

Now we prove that LP(𝔼[G])𝔼[LP(G)]LP(\operatorname{\mathbb{E}_{\mathcal{I}}}[G])\geq\operatorname{\mathbb{E}_{\mathcal{I}}}[LP(G)]. The dual of the LP for the realization GG^{\prime} is the following:

minuU(αu+Δuβu)+vV(αv+Δvβv)+(u,v)γu,v\displaystyle\textstyle\min\sum_{u\in U}(\alpha_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\beta_{u})+\sum_{v^{\prime}\in V^{\prime}}(\alpha_{v^{\prime}}+\Delta_{v^{\prime}}\beta_{v^{\prime}})+\sum_{(u,v^{\prime})}\gamma_{u,v^{\prime}} (11)
s.t.uU,vV:\displaystyle\text{s.t.}\quad\forall u\in U,\forall v^{\prime}\in V^{\prime}:
βu+βv+p(u,v)(αu+αv)+γ(u,v)w(u,v)Op(u,v)\displaystyle\beta_{u}+\beta_{v^{\prime}}+p_{(u,v^{\prime})}(\alpha_{u}+\alpha_{v^{\prime}})+\gamma_{(u,v^{\prime})}\geq w^{O}_{(u,v^{\prime})}p_{(u,v^{\prime})} (12a)
αu,αv,βu,βv,γ(u,v)0\displaystyle\alpha_{u},\alpha_{v^{\prime}},\beta_{u},\beta_{v^{\prime}},\gamma_{(u,v^{\prime})}\geq 0 (12b)

Consider the graph with the expected number of arrival 𝔼(G)\operatorname{\mathbb{E}_{\mathcal{I}}}(G) it would have a dual of the above form, let α,β,γ\vec{\alpha}^{*},\vec{\beta}^{*},\vec{\gamma}^{*} be the optimal solution of its corresponding dual. Then it follows by the strong duality of LPs that solution α,β,γ\vec{\alpha}^{*},\vec{\beta}^{*},\vec{\gamma}^{*} would have a value of LP(𝔼[G])LP(\operatorname{\mathbb{E}_{\mathcal{I}}}[G]). Now for the instance GG^{\prime}, we shall use the following dual solution α^,β^,γ^\vec{\mathnormal{\hat{\alpha}}},\vec{\mathnormal{\hat{\beta}}},\vec{\mathnormal{\hat{\gamma}}} which is set as follows:

  • uU:α^u=αu,β^u=αu\forall u\in U:\mathnormal{\hat{\alpha}}_{u}=\alpha^{*}_{u},\mathnormal{\hat{\beta}}_{u}=\alpha^{*}_{u}.

  • vV\forall v^{\prime}\in V^{\prime} of type vv: α^v=αv,β^v=βv\mathnormal{\hat{\alpha}}_{v^{\prime}}=\alpha^{*}_{v},\mathnormal{\hat{\beta}}_{v^{\prime}}=\beta^{*}_{v}.

  • uU,vV\forall u\in U,\forall v^{\prime}\in V^{\prime} of type vv: γ^(u,v)=γ(u,v)\mathnormal{\hat{\gamma}}_{(u,v^{\prime})}=\gamma^{*}_{(u,v)}.

Note that the new solution α^,β^,γ^\vec{\mathnormal{\hat{\alpha}}},\vec{\mathnormal{\hat{\beta}}},\vec{\mathnormal{\hat{\gamma}}} is a feasible dual solution since it satisfies constraints 12a and 12b. By weak duality the value of the solution α^,β^,γ^\vec{\mathnormal{\hat{\alpha}}},\vec{\mathnormal{\hat{\beta}}},\vec{\mathnormal{\hat{\gamma}}} upper bounds LP(G)LP(G^{\prime}). Now if we were to denote the number of vertices of type vv that arrived in instance GG^{\prime} by nvGn^{G^{\prime}}_{v}, then the value of the solution α^,β^,γ^\vec{\mathnormal{\hat{\alpha}}},\vec{\mathnormal{\hat{\beta}}},\vec{\mathnormal{\hat{\gamma}}} satisfies:

uU(α^u+Δuβ^u)+vV(α^v+Δvβ^v)+(u,v)γ^u,v\displaystyle\sum_{u\in U}(\mathnormal{\hat{\alpha}}_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\mathnormal{\hat{\beta}}_{u})+\sum_{v^{\prime}\in V^{\prime}}(\mathnormal{\hat{\alpha}}_{v^{\prime}}+\Delta_{v^{\prime}}\mathnormal{\hat{\beta}}_{v^{\prime}})+\sum_{(u,v^{\prime})}\mathnormal{\hat{\gamma}}_{u,v^{\prime}}
=uU(αu+Δuβu)+vVnvG(αv+Δvβv)+(u,v)nvGγu,v\displaystyle=\sum_{u\in U}(\alpha^{*}_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\beta^{*}_{u})+\sum_{v\in V}n^{G^{\prime}}_{v}(\alpha^{*}_{v}+\Delta_{v}\beta^{*}_{v})+\sum_{(u,v)}n^{G^{\prime}}_{v}\gamma^{*}_{u,v}
LP(G)\displaystyle\geq LP(G^{\prime})

Now taking the expectation, we get:

𝔼[LP(G)]\displaystyle\operatorname{\mathbb{E}_{\mathcal{I}}}[LP(G^{\prime})]
𝔼[uU(α^u+Δuβ^u)+vV(α^v+Δvβ^v)+(u,v)γ^u,v]\displaystyle\leq\operatorname{\mathbb{E}_{\mathcal{I}}}\Big{[}\sum_{u\in U}(\mathnormal{\hat{\alpha}}_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\mathnormal{\hat{\beta}}_{u})+\sum_{v^{\prime}\in V^{\prime}}(\mathnormal{\hat{\alpha}}_{v^{\prime}}+\Delta_{v^{\prime}}\mathnormal{\hat{\beta}}_{v^{\prime}})+\sum_{(u,v^{\prime})}\mathnormal{\hat{\gamma}}_{u,v^{\prime}}\Big{]}
=𝔼[uU(αu+Δuβu)+vVnvG(αv+Δvβv)+(u,v)nvGγu,v]\displaystyle=\operatorname{\mathbb{E}_{\mathcal{I}}}\Big{[}\sum_{u\in U}(\alpha^{*}_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\beta^{*}_{u})+\sum_{v\in V}n^{G^{\prime}}_{v}(\alpha^{*}_{v}+\Delta_{v}\beta^{*}_{v})+\sum_{(u,v)}n^{G^{\prime}}_{v}\gamma^{*}_{u,v}\Big{]}
=uU(αu+Δuβu)+vV𝔼[nvG](αv+Δvβv)+(u,v)𝔼[nvG]γu,v\displaystyle=\sum_{u\in U}(\alpha^{*}_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\beta^{*}_{u})+\sum_{v\in V}\operatorname{\mathbb{E}_{\mathcal{I}}}[n^{G^{\prime}}_{v}](\alpha^{*}_{v}+\Delta_{v}\beta^{*}_{v})+\sum_{(u,v)}\operatorname{\mathbb{E}_{\mathcal{I}}}[n^{G^{\prime}}_{v}]\gamma^{*}_{u,v}
=uU(αu+Δuβu)+vV(αv+Δvβv)+(u,v)γu,v\displaystyle=\sum_{u\in U}(\alpha^{*}_{u}+\mathnormal{\Delta_{\mathnormal{u}}}\beta^{*}_{u})+\sum_{v\in V}(\alpha^{*}_{v}+\Delta_{v}\beta^{*}_{v})+\sum_{(u,v)}\gamma^{*}_{u,v}
=LP(𝔼[G])\displaystyle=LP(\operatorname{\mathbb{E}_{\mathcal{I}}}[G])

For the offline and online group fairness objectives, we use the same steps. The difference would be in the constraints of the dual program, however following a similar assignment as done from α,β,γ\vec{\alpha}^{*},\vec{\beta}^{*},\vec{\gamma}^{*} to α^,β^,γ^\vec{\mathnormal{\hat{\alpha}}},\vec{\mathnormal{\hat{\beta}}},\vec{\mathnormal{\hat{\gamma}}} is sufficient to prove the lemma for both fairness objectives. ∎

Before we prove Lemma 5.2 for the lower bound on the probability of SFu,tSF_{u,t}. We have to first introduce the following two lemmas. Specifically, let Au,tA_{u,t} be the number of successful assignments that uu received and accepted before round tt. Then the following lemma holds.

Lemma B.1.

For any given vertex uu at time t[T]t\in[T] , P[Au,t=0](11T)t1P[A_{u,t}=0]\geq\Big{(}1-\frac{1}{T}\Big{)}^{t-1}.

Proof.

Let Xe,kX_{e,k} be the indicator random variable for uu receiving an arrival request of type vv where eEue\in E_{u} and k<tk<t. Let Ye,kY_{e,k} be the indicator random variable that the edge ee gets sampled by the TSGFKIID(α,β,γ)\operatorname{\mathrm{TSGF}_{\textbf{KIID}}}(\alpha,\beta,\gamma) algorithm at time k<tk<t. Let Ze,kZ_{e,k} be the indicator random variable that assignment e=(u,v)e=(u,v) is successful (a match) at time k<tk<t. Then Au,t=k<teEuXe,kYe,kZe,kA_{u,t}=\sum_{k<t}\sum_{e\in E_{u}}X_{e,k}Y_{e,k}Z_{e,k}.

Pr[Au,t=0]=Πk<tPr[e=(u,v)EuXe,kYe,kZe,k=0]\displaystyle Pr[A_{u,t}=0]=\Pi_{k<t}Pr\Big{[}\sum_{e=(u,v)\in E_{u}}X_{e,k}Y_{e,k}Z_{e,k}=0\Big{]}
=Πk<t(1Pr[eEuXe,kYe,kZe,k1])\displaystyle=\Pi_{k<t}\Big{(}1-Pr\Big{[}\sum_{e\in E_{u}}X_{e,k}Y_{e,k}Z_{e,k}\geq 1\Big{]}\Big{)}
Πk<t(1eEu1T(αxe+βyeqv+γzeqv)pe)\displaystyle\geq\Pi_{k<t}\Big{(}1-\sum_{e\in E_{u}}\frac{1}{T}\cdot\big{(}\alpha x_{e}^{*}+\beta\frac{y_{e}^{*}}{q_{v}}+\gamma\frac{z_{e}^{*}}{q_{v}}\big{)}\cdot p_{e}\Big{)}
=Πk<t(11T(αeEuxepe+βeEuyepe+γeEuzepe))\displaystyle=\Pi_{k<t}\Big{(}1-\frac{1}{T}\cdot\big{(}\alpha\sum_{e\in E_{u}}{x_{e}^{*}p_{e}}+\beta\sum_{e\in E_{u}}{y_{e}^{*}p_{e}}+\gamma\sum_{e\in E_{u}}{z_{e}^{*}p_{e}}\big{)}\Big{)}
Πk<t(11T(α1+β1+γ1))\displaystyle\geq\Pi_{k<t}\Big{(}1-\frac{1}{T}\cdot\big{(}\alpha\cdot 1+\beta\cdot 1+\gamma\cdot 1\big{)}\Big{)}
Πk<t(11T)=(11T)t1\displaystyle\geq\Pi_{k<t}\Big{(}1-\frac{1}{T}\Big{)}=\Big{(}1-\frac{1}{T}\Big{)}^{t-1}

Now we lower bound the probability that uu was probed less than Δu\Delta_{u} times prior to tt. Denote the number of probes received by uu before tt by Bu,tB_{u,t}, then the following lemma holds:

Lemma B.2.

Pr[Bu,t<Δu]1t1TPr[B_{u,t}<\Delta_{u}]\geq 1-\frac{t-1}{T}.

Proof.

First it is clear that Bu,t=k<teEuXe,kYe,kB_{u,t}=\sum_{k<t}\sum_{e\in E_{u}}X_{e,k}Y_{e,k}.

𝔼[Bu,t]=k<teEu𝔼[Xe,kYe,k]\displaystyle\mathbb{E}[B_{u,t}]=\sum_{k<t}\sum_{e\in E_{u}}\mathbb{E}[X_{e,k}Y_{e,k}]
k<teEu1T(αxe+βye+γze)\displaystyle\leq\sum_{k<t}\sum_{e\in E_{u}}\frac{1}{T}\Big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\Big{)}
k<t1T(αeEdxe+βeEuye+γeEuze)\displaystyle\leq\sum_{k<t}\frac{1}{T}\Big{(}\alpha\sum_{e\in E_{d}}{x_{e}^{*}}+\beta\sum_{e\in E_{u}}{y_{e}^{*}}+\gamma\sum_{e\in E_{u}}z_{e}^{*}\Big{)}
k<tΔuT(α+β+γ)(t1)ΔuT\displaystyle\leq\sum_{k<t}\frac{\Delta_{u}}{T}(\alpha+\beta+\gamma)\leq\frac{(t-1)\Delta_{u}}{T}

The inequality before the last follows from (α+β+γ)1(\alpha+\beta+\gamma)\leq 1. Now using Markov’s inequality: Pr[Bu,t<Δu]1𝔼[Bu,t]ΔuPr[B_{u,t}<\Delta_{u}]\geq 1-\frac{\mathbb{E}[B_{u,t}]}{\Delta_{u}}, we get Pr[Bu,t<Δu]1t1T\implies Pr[B_{u,t}<\Delta_{u}]\geq 1-\frac{t-1}{T}. ∎

Now we restate Lemma 5.2 and prove it. See 5.2

Proof.

Consider a given edge eEue\in E_{u} where k<tk<t

𝔼[Xe,kYe,kAu,t=0]=𝔼[Xe,kYe,kAu,k=0]\displaystyle\mathbb{E}[X_{e,k}Y_{e,k}\mid A_{u,t}=0]=\mathbb{E}[X_{e,k}Y_{e,k}\mid A_{u,k}=0]
=Pr[Xe,k=1,Ye,k=1,Ze,k=0]Pr[Au,k=0]\displaystyle=\frac{Pr[X_{e,k}=1,Y_{e,k}=1,Z_{e,k}=0]}{Pr[A_{u,k}=0]}
1T(αxe+βye+γze)(1pe)1eEd1T(αxe+βye+γze)pe\displaystyle\leq\frac{\frac{1}{T}\cdot\big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\big{)}\cdot(1-p_{e})}{1-\sum_{e\in E_{d}}\frac{1}{T}\cdot\big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\big{)}\cdot p_{e}}
=1T(αxe+βye+γze)(1pe)1pe+pe(1eEd1T(αxe+βye+γze))\displaystyle=\frac{\frac{1}{T}\cdot\big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\big{)}\cdot(1-p_{e})}{1-p_{e}+p_{e}\Big{(}1-\sum_{e\in E_{d}}\frac{1}{T}\cdot\big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\big{)}\Big{)}}
1T(αxe+βye+γze)\displaystyle\leq{\frac{1}{T}\cdot\big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\big{)}\cdot}

The above inequality is due to the fact that eEu1T(αxe+βye+γze)ΔuT<1\sum_{e\in E_{u}}\frac{1}{T}\big{(}\alpha{x_{e}^{*}}+\beta{y_{e}^{*}}+\gamma{z_{e}^{*}}\big{)}\leq\frac{\mathnormal{\Delta_{\mathnormal{u}}}}{T}<1.

𝔼[Bu,t|Au,t=0]=k<teEu𝔼[Xe,kYe,k|Au,k=0]\displaystyle\mathbb{E}[B_{u,t}|A_{u,t}=0]=\sum_{k<t}\sum_{e\in E_{u}}\mathbb{E}[X_{e,k}Y_{e,k}|A_{u,k}=0]
k<teEu1T(αxe+βye+γze)\displaystyle\leq\sum_{k<t}\sum_{e\in E_{u}}\frac{1}{T}\Big{(}\alpha x_{e}^{*}+\beta y_{e}^{*}+\gamma z_{e}^{*}\Big{)}
k<t1T(αeEuxe+βeEuye+γeEuze)\displaystyle\leq\sum_{k<t}\frac{1}{T}\Big{(}\alpha\sum_{e\in E_{u}}{x_{e}^{*}}+\beta\sum_{e\in E_{u}}{y_{e}^{*}}+\gamma\sum_{e\in E_{u}}z_{e}^{*}\Big{)}
k<t1T(αΔu+βΔd+γΔu)\displaystyle\leq\sum_{k<t}\frac{1}{T}\Big{(}\alpha\cdot\Delta_{u}+\beta\cdot\Delta_{d}+\gamma\cdot\Delta_{u}\Big{)}
=k<tΔuT(α+β+γ)(t1)ΔuT\displaystyle=\sum_{k<t}\frac{\Delta_{u}}{T}(\alpha+\beta+\gamma)\leq\frac{(t-1)\Delta_{u}}{T}

Therefore the expected number of assignments (probes) to vertex uu until time tt is at most (t1)ΔuT\frac{(t-1)\Delta_{u}}{T}. Therefore, we have:

Pr[Bu,t<Δu|Au,t=0]1𝔼[Bu,t|Au,t=0]Δd\displaystyle Pr[B_{u,t}<\Delta_{u}|A_{u,t}=0]\geq 1-\frac{\mathbb{E}[B_{u,t}|A_{u,t}=0]}{\Delta_{d}}
1(t1)ΔuTΔu1t1T\displaystyle\geq 1-\frac{(t-1)\Delta_{u}}{T\Delta_{u}}\geq 1-\frac{t-1}{T}

It is to be noted that Bu,tB_{u,t} is the total number of probes uu received before round tt. Thus, we have that the events (Bu,t<Δu)(B_{u,t}<\Delta_{u}) and (Au,t=0)(A_{u,t}=0) are positively correlated. Therefore,

Pr[SFu,t]Pr[(Bu,t<Δu)(Au,t=0)]\displaystyle Pr[SF_{u,t}]\geq Pr[(B_{u,t}<\Delta_{u})\land(A_{u,t}=0)]
Pr[Bu,t<Δd|Au,t=0]Pr[Au,t=0]\displaystyle\geq Pr[B_{u,t}<\Delta_{d}|A_{u,t}=0]Pr[A_{u,t}=0]
Pr[SFu,t](1t1T)(11T)t1\displaystyle Pr[SF_{u,t}]\geq\Big{(}1-\frac{t-1}{T}\Big{)}\Big{(}1-\frac{1}{T}\Big{)}^{t-1}

. See 5.3

Proof.

In this part we prove that the probability that edge ee is probed at time tt is at least αxe2T\alpha\frac{x^{*}_{e}}{2T}. Note that the probability that a vertex of type vv arrives at time tt and that Algorithm 2 calls the subroutine PPDR(xr\vec{x_{r}}) is α1T\alpha\frac{1}{T}. Let Ev,e¯E_{v,\bar{e}} be the set of edges in EvE_{v} excluding e=(u,v)e=(u,v). For each edge eEv,e¯e^{\prime}\in E_{v,\bar{e}} let YeY_{e^{\prime}} be the indicator for ee^{\prime} being before ee in the random order of π\pi (in algorithm 1) and let ZeZ_{e^{\prime}} be the probability that the assignment is successful for ee^{\prime}. It is clear that 𝔼[Ye]=1/2\operatorname{\mathbb{E}}[Y_{e^{\prime}}]=1/2 and that 𝔼[Ze]=pe\operatorname{\mathbb{E}}[Z_{e^{\prime}}]=p_{e^{\prime}}. Now we have:

Pr[1e,tSFu,t]\displaystyle Pr[1_{e,t}\mid SF_{u,t}] (13)
α1TPr[Xe=1]Pr[eEr,e¯XeYeZeXe=1]\displaystyle\geq\alpha\frac{1}{T}Pr[X_{e}=1]Pr\big{[}\sum_{e^{\prime}\in E_{r,\bar{e}}}X_{e^{\prime}}Y_{e^{\prime}}Z_{e^{\prime}}\mid X_{e}=1\big{]} (14)
=αPr[Xe=1]T(1Pr[eEv,e¯XeYeZe1Xe=1])\displaystyle=\alpha\frac{Pr[X_{e}=1]}{T}\big{(}1-Pr\big{[}\sum_{e^{\prime}\in E_{v,\bar{e}}}X_{e^{\prime}}Y_{e^{\prime}}Z_{e^{\prime}}\geq 1\mid X_{e}=1\big{]}\big{)} (15)
αPr[Xe=1]T(1𝔼[eEv,e¯XeYeZe1Xe=1])\displaystyle\geq\alpha\frac{Pr[X_{e}=1]}{T}\big{(}1-\mathbb{E}\big{[}\sum_{e^{\prime}\in E_{v,\bar{e}}}X_{e^{\prime}}Y_{e^{\prime}}Z_{e^{\prime}}\geq 1\mid X_{e}=1\big{]}\big{)} (16)
αPr[Xe=1]T(1eEv,e¯𝔼[XeYeZe1Xe=1])\displaystyle\geq\alpha\frac{Pr[X_{e}=1]}{T}\big{(}1-\sum_{e^{\prime}\in E_{v,\bar{e}}}\mathbb{E}\big{[}X_{e^{\prime}}Y_{e^{\prime}}Z_{e^{\prime}}\geq 1\mid X_{e}=1\big{]}\big{)} (17)
αxeT(1eEv,e¯xe12pe)\displaystyle\geq\alpha\frac{x^{*}_{e}}{T}\big{(}1-\sum_{e^{\prime}\in E_{v,\bar{e}}}x_{e^{\prime}}^{*}\frac{1}{2}p_{e^{\prime}}\big{)} (18)
αxeT(112)=αxe2T\displaystyle\geq\alpha\frac{x^{*}_{e}}{T}\big{(}1-\frac{1}{2}\big{)}=\alpha\frac{x^{*}_{e}}{2T} (19)

Applying Markov inequality we get the inequality (16). By linearity of expectation we get inequality (17). Since XeX_{e} and XeX_{e^{\prime}} are negatively correlated to each other from the Negative Correlation property of Dependent Rounding we have 𝔼[XeXe=1]xe\operatorname{\mathbb{E}}[X_{e^{\prime}}\mid X_{e}=1]\leq{x}_{e}^{*} and we get (18). The last inequality (19) is due the fact that for any feasible solution {xe}\{x_{e}^{*}\} the constraints imply that eEvxepe1\sum_{e\in E_{v}}x_{e}^{*}p_{e}\leq 1 for all vVv\in V. Using similar analysis we can also prove that Pr[1e,tSFu,t]βye2TPr[1_{e,t}\mid SF_{u,t}]\geq\beta\frac{y^{*}_{e}}{2T} and Pr[1e,tSFu,t]γze2TPr[1_{e,t}\mid SF_{u,t}]\geq\gamma\frac{z^{*}_{e}}{2T}. ∎

Now we restate and prove Theorem 4.1. See 4.1

Proof.

Denote the expected number of probes on each edge eEe\in E resulting from PPDR(xv\vec{x}_{v}^{*}) by nexn^{x}_{e}. It follows that:

next=1TPr[1e,t]=t=1TPr[1e,tSFu,t]Pr[SFu,t]\displaystyle n^{x}_{e}\geq\sum_{t=1}^{T}Pr[1_{e,t}]=\sum_{t=1}^{T}Pr[1_{e,t}\mid SF_{u,t}]Pr[SF_{u,t}]
t=1T(11T)t1(1t1T)(αxe2T)Tαxe2e\displaystyle\geq\sum_{t=1}^{T}\Big{(}1-\frac{1}{T}\Big{)}^{t-1}\Big{(}1-\frac{t-1}{T}\Big{)}\Big{(}\alpha\frac{x^{*}_{e}}{2T}\Big{)}\xrightarrow{T\rightarrow\infty}\frac{\alpha x^{*}_{e}}{2e}

Denote the optimal solution for the operator’s profit LP by OPTOOPT_{O}. Let ALGOALG_{O} be operator’s profit obtained by our online algorithm. Using the linearity of expectation we get: ALGO=𝔼[eEweOnexpe]eEweOpeαxe2eeEweOpe(1e)αxe2α2e(OPTO){ALG_{O}}=\mathbb{E}\Big{[}\sum_{e\in E}{\mathnormal{w^{O}_{e}}n^{x}_{e}p_{e}}\Big{]}\geq\sum_{e\in E}{\mathnormal{w^{O}_{e}}p_{e}}\frac{\alpha x^{*}_{e}}{2e}\geq\sum_{e\in E}{\mathnormal{w^{O}_{e}}p_{e}}\Big{(}\frac{1}{e}\Big{)}\frac{\alpha x^{*}_{e}}{2}\geq\frac{\alpha}{2e}(OPT_{O}). Similarly, we can obtain β2e\frac{\beta}{2e} and γ2e\frac{\gamma}{2e} competitive ratios for the expected max-min group fairness guarantees on the offline and online sides, respectively. ∎

B.2 Proofs for Section 5.2

See 5.4

Proof.

We shall consider only the operator’s profit objective as the other objectives follow through an identical argument. Let 1v,t1_{v,t} be the indicator random variable for the arrival for vertex type vv in round tt. Then we can obtain a realization and solve the corresponding LP and then take the expected value of LP as an upper bound on the operator’s profit objective, i.e. the value 𝔼[LP(G)]\operatorname{\mathbb{E}}_{\mathcal{I}}[LP(G)] where 𝔼\operatorname{\mathbb{E}}_{\mathcal{I}} is an expectation with respect to the randomness of the problem. This means replacing 1v,t1_{v,t} by its realization in the LP below:

maxt[T]eEweOxe,t\displaystyle\textstyle\max\sum\limits_{t\in[T]}\sum\limits_{e\in E}{\mathnormal{w^{O}_{e}}\mathnormal{x_{e,t}}} (20)
s.teE,t[T]:0xe,t1\displaystyle\text{s.t}\quad\forall e\in E,\forall t\in[T]:0\leq\mathnormal{x_{e,t}}\leq 1 (21a)
uU:t[T]eEuxe,t1\displaystyle\textstyle\forall u\in U:\sum\limits_{t\in[T]}\sum\limits_{e\in E_{u}}\mathnormal{x_{e,t}}\leq 1 (21b)
vV,t[T]:eEvxe,t1v,t\displaystyle\textstyle\forall v\in V,\forall t\in[T]:\sum_{e\in E_{v}}\mathnormal{x_{e,t}}\leq 1_{v,t} (21c)

If we were to replace the random variables 1v,t1_{v,t} by their expected value, then we would retrieve LP(5) where 𝔼[1v,t]=pv,t\operatorname{\mathbb{E}}_{\mathcal{I}}[1_{v,t}]=\mathnormal{p_{v,t}}. It suffices to show that the value of LP(5) which is the LP value over the “expected” graph (the parameters replaced by their expected value) which we now denote by LP(𝔼[G])LP(\operatorname{\mathbb{E}}_{\mathcal{I}}[G]) is an upper bound to 𝔼[LP(G)]\operatorname{\mathbb{E}}_{\mathcal{I}}[LP(G)], i.e. LP(𝔼[G])𝔼[LP(G)]LP(\operatorname{\mathbb{E}}_{\mathcal{I}}[G])\geq\operatorname{\mathbb{E}}_{\mathcal{I}}[LP(G)]. Let xe,t,Gx^{*,G}_{e,t} be the optimal solution for a given realization GG and 1v,tG1^{G}_{v,t} be the realization of the random variables over the instance, then we have that eEvxe,t,G1v,tG\sum_{e\in E_{v}}x^{*,G}_{e,t}\leq 1^{G}_{v,t}. It follows that 𝔼[xe,t,G]\operatorname{\mathbb{E}}_{\mathcal{I}}[x^{*,G}_{e,t}] is a feasible solution for LP(5), since 𝔼[eEvxe,t,G]𝔼[1v,tG]=pv,t\operatorname{\mathbb{E}}_{\mathcal{I}}[\sum_{e\in E_{v}}x^{*,G}_{e,t}]\leq\operatorname{\mathbb{E}}_{\mathcal{I}}[1^{G}_{v,t}]=\mathnormal{p_{v,t}} and the rest of the constraints are satisfied as well since they are the same in every realization. However, we have that 𝔼[LP(G)]=𝔼[t[T]eEweOxe,t,G]=t[T]eEweO𝔼[xe,t,G]t[T]eEweOxe,t=LP(𝔼[G])\operatorname{\mathbb{E}}_{\mathcal{I}}[LP(G)]=\operatorname{\mathbb{E}}_{\mathcal{I}}[\sum\limits_{t\in[T]}\sum\limits_{e\in E}{\mathnormal{w^{O}_{e}}x^{*,G}_{e,t}}]=\sum\limits_{t\in[T]}\sum\limits_{e\in E}{\mathnormal{w^{O}_{e}}\operatorname{\mathbb{E}}_{\mathcal{I}}[x^{*,G}_{e,t}]}\leq\sum\limits_{t\in[T]}\sum\limits_{e\in E}{\mathnormal{w^{O}_{e}}x^{*}_{e,t}}=LP(\operatorname{\mathbb{E}}_{\mathcal{I}}[G]) where xe,tx^{*}_{e,t} is the optimal solution for LP(5) over the “expected” graph. The inequality followed since a feasible solution to a problem cannot exceed its optimal solution. ∎

See 5.5

Proof.

We prove the validity of the algorithm for λ=12\lambda=\frac{1}{2} by induction. For the base case, it is clear that eE,ρe,t=1\forall e\in E,\mathnormal{\rho_{e,t}}=1, hence ρe,tλ=12\mathnormal{\rho_{e,t}}\geq\lambda=\frac{1}{2}. Assume for t<tt^{\prime}<t, that ρe,tλ=12\rho_{e,t^{\prime}}\geq\lambda=\frac{1}{2}, then at round tt we have:

1ρe,t\displaystyle 1-\mathnormal{\rho_{e,t}} =Pr[e is not available at t]\displaystyle=\Pr[\text{$e$ is not available at $t$}]
=Pr[e is matched in [T1]]\displaystyle=\Pr[\text{$e$ is matched in $[T-1]$}]
t<tPr[e is matched in t]\displaystyle\leq\sum_{t^{\prime}<t}\Pr[\text{$e$ is matched in $t^{\prime}$}]
=t<tPr[(e is chosen by the algorithm)\displaystyle=\sum_{t^{\prime}<t}\Pr[(\text{$e$ is chosen by the algorithm})
(u is unmatched at the beginning of t)\displaystyle\land(\text{$u$ is unmatched at the beginning of $t$})
(v arrives at t)]\displaystyle\land(\text{$v$ arrives at $t$})]
=t<tpv,tρe,t(αxe,tpv,tλρe,t+βye,tpv,tλρe,t+γze,tpv,tλρe,t)\displaystyle=\sum_{t^{\prime}<t}\mathnormal{p_{v,t}}\mathnormal{\rho_{e,t}}(\alpha\frac{\mathnormal{x^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\beta\frac{\mathnormal{y^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\gamma\frac{\mathnormal{z^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}})
=t<tλ(αxe,t+βye,t+γze,t)\displaystyle=\sum_{t^{\prime}<t}\lambda(\alpha x^{*}_{e,t^{\prime}}+\beta y^{*}_{e,t^{\prime}}+\gamma z^{*}_{e,t^{\prime}})
λt<t(αxe,t+βye,t+γze,t)\displaystyle\leq\lambda\sum_{t^{\prime}<t}(\alpha x^{*}_{e,t^{\prime}}+\beta y^{*}_{e,t^{\prime}}+\gamma z^{*}_{e,t^{\prime}})
λ(α+β+γ)λ12\displaystyle\leq\lambda(\alpha+\beta+\gamma)\leq\lambda\leq\frac{1}{2}

where we used the fact that xe,t,ye,t,ze,t1x^{*}_{e,t^{\prime}},y^{*}_{e,t^{\prime}},z^{*}_{e,t^{\prime}}\leq 1 from constraint (8a) and the fact that α+β+γ1\alpha+\beta+\gamma\leq 1. From the above, it follows that ρe,t12λ\mathnormal{\rho_{e,t}}\geq\frac{1}{2}\geq\lambda. ∎

Now we restate and prove Theorem 4.2 using Lemma 5.5: See 4.2

Proof.

For an edge ee the probability that it is matched (successfully probed) is the following:

Pr[e is successfully probed in round t]\displaystyle\Pr[\text{$e$ is successfully probed in round $t$}]
=Pr[(e is chosen by the algorithm)\displaystyle=\Pr[(\text{$e$ is chosen by the algorithm})
(u is unmatched at the beginning of t)(v arrives at t)]\displaystyle\land(\text{$u$ is unmatched at the beginning of $t$})\land(\text{$v$ arrives at $t$})]
=pv,tρe,t(αxe,tpv,tλρe,t+βye,tpv,tλρe,t+γze,tpv,tλρe,t)=\displaystyle=\mathnormal{p_{v,t}}\mathnormal{\rho_{e,t}}(\alpha\frac{\mathnormal{x^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\beta\frac{\mathnormal{y^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}}+\gamma\frac{\mathnormal{z^{*}_{e,t}}}{\mathnormal{p_{v,t}}}\frac{\lambda}{\mathnormal{\rho_{e,t}}})=
=αλxe,t+βλye,t+γλze,t\displaystyle=\alpha\lambda\mathnormal{x^{*}_{e,t}}+\beta\lambda\mathnormal{y^{*}_{e,t}}+\gamma\lambda\mathnormal{z^{*}_{e,t}}

Setting λ=12\lambda=\frac{1}{2}, it follows from the above that ee is successfully matched with probability at least 12αxe,t\frac{1}{2}\alpha\mathnormal{x^{*}_{e,t}}, at least 12βye,t\frac{1}{2}\beta\mathnormal{y^{*}_{e,t}}, and at least 12γze,t\frac{1}{2}\gamma\mathnormal{z^{*}_{e,t}}. Hence, the guarantees on the competitive ratios follow by linearity of the expectation. ∎

B.3 Proofs for Section 5.3

We restate Lemma 5.6 and give its proof: See 5.6

Proof.

Given an instance with individual fairness, define 𝒢={g1,,gT}{g1,,g|U|}\mathcal{G}=\{g_{1},\dots,g_{T}\}\cup\{g^{\prime}_{1},\dots,g^{\prime}_{|U|}\} as the set of all groups, thus |𝒢|=T+|U||\mathcal{G}|=T+|U|, i.e. one group for each time round and one group for each offline vertex. Further given the online side types VV, create a new online side VV^{\prime} where |V|=T|V||V^{\prime}|=T|V| and V=V1V2VtVTV^{\prime}=V^{\prime}_{1}\cup V^{\prime}_{2}\dots\cup V^{\prime}_{t}\dots\cup V^{\prime}_{T} where VtV^{\prime}_{t} consists of the same types as VV. Moreover, vVt,pv,t=pv,t\forall v^{\prime}\in V^{\prime}_{t},p_{v^{\prime},t}=\mathnormal{p_{v,t}} and pv,t¯=0,t¯[T]{t}p_{v^{\prime},\bar{t}}=0,\forall\bar{t}\in[T]-\{t\}, finally vVt,g(v)=gt\forall v^{\prime}\in V^{\prime}_{t},g(v^{\prime})=g_{t}. For the offline side UU, we let each vertex have its own distinct group membership, i.e. for vertex uiUu_{i}\in U, g(ui)=gig(u_{i})=g^{\prime}_{i}.

Based on the above, it is not difficult to see that both problems have the same operator profit, and that the individual max-min fairness objectives of the original instance equal the group max-min fairness objectives of the new instance. ∎

From the above Lemma, applying algorithm TSGFKAD\operatorname{\mathrm{TSGF}_{\textbf{KAD}}} to the reduced instance leads to the following corollary:

Corollary B.1.

Given an instance of two-sided individual max-min fairness, applying TSGFKAD(α,β,γ)\operatorname{\mathrm{TSGF}_{\textbf{KAD}}}(\alpha,\beta,\gamma) to the reduction from Theorem 5.6 leads to a competitive ratio of (α2,β2,γ2)(\frac{\alpha}{2},\frac{\beta}{2},\frac{\gamma}{2}) simultaneously over the operator’s profit, the individual fairness objective for the offline side, and the individual fairness objective for the online side, where α,β,γ>0\alpha,\beta,\gamma>0 and α+β+γ1\alpha+\beta+\gamma\leq 1.

The proof of Theorem 4.3 is immediate from the above corollary.

B.4 Proofs for Theorems 4.4 and 4.5

We now restate and prove the hardness result of Theorem 4.4: See 4.4

Proof.

We prove it for group fairness in the KIID setting, since the KIID setting is a special case of the KAD setting, then this also proves the upper bound for the KAD setting.

Consider the graph G=(U,V,E)G=(U,V,E) which consists of three offline vertices and three online vertex types, i.e. |U|=|V|=3|U|=|V|=3. Each vertex in UU (VV) belongs to its own distinct group. The time horizon TT is set to an arbitrarily large value. The arrival rate for each vVv\in V is uniform and independent of time, i.e. KIID with pv=13\mathnormal{p_{v}}=\frac{1}{3}. Further, the bipartite graph is complete, i.e. each vertex of UU is connected to all of the vertices of VV with pe=1p_{e}=1 for all eEe\in E. We also let Δu=1\Delta_{u}=1 for each uUu\in U, nv=T3n_{v}=\frac{T}{3} and Δv=1\Delta_{v}=1 for each vVv\in V. We represent the utilities on the edges of EE with matrices where the (i,j)(i,j) element gives the utility of the edge connecting vertex uiUu_{i}\in U and vertex vjVv_{j}\in V. The utility matrices for the platform operator, offline, and online sides are following, respectively:

MO=[100010001],MU=[001100010],MV=[010001100].M_{O}=\begin{bmatrix}1&0&0\\ 0&1&0\\ 0&0&1\end{bmatrix},M_{U}=\begin{bmatrix}0&0&1\\ 1&0&0\\ 0&1&0\end{bmatrix},M_{V}=\begin{bmatrix}0&1&0\\ 0&0&1\\ 1&0&0\end{bmatrix}.

It can be seen that the utility assignments in the above example conflict between the three entities.

Let OPTO,OPTU,\operatorname{\mathrm{OPT}}_{O},\operatorname{\mathrm{OPT}}_{U}, and OPTV\operatorname{\mathrm{OPT}}_{V} be the optimal values for the operator’s profit, offline group fairness, and online group fairness, respectively. It is not difficult to see that OPTO=3\operatorname{\mathrm{OPT}}_{O}=3, OPTU=1\operatorname{\mathrm{OPT}}_{U}=1, and OPTV=1\operatorname{\mathrm{OPT}}_{V}=1. Now, denote by A,B,A,B, and CC the edges with values of 1 for MO,MU,M_{O},M_{U}, and MVM_{V} in the graph, respectively. Further, for a given online algorithm, let aj,bk,a_{j},b_{k}, and cc_{\ell} be the expected number of probes received by edges jA,kB,j\in A,k\in B, and C\ell\in C, respectively. Moreover, denote the algorithm’s expected value over the operator’s profit, expected fairness for offline vertices, and expected fairness for online vertices by ALGO,ALGU\operatorname{\mathrm{ALG}}_{O},\operatorname{\mathrm{ALG}}_{U}, and ALGV\operatorname{\mathrm{ALG}}_{V}, respectively. We can upper bound the sum of the competitive ratios as follows:

ALGOOPTO+ALGUOPTU+ALGVOPTV\displaystyle\frac{\operatorname{\mathrm{ALG}}_{O}}{\operatorname{\mathrm{OPT}}_{O}}+\frac{\operatorname{\mathrm{ALG}}_{U}}{\operatorname{\mathrm{OPT}}_{U}}+\frac{ALG_{V}}{\operatorname{\mathrm{OPT}}_{V}}
jAaj3+minkBbj1+minCcj1\displaystyle\leq\frac{\sum_{j\in A}a_{j}}{3}+\frac{\min_{k\in B}b_{j}}{1}+\frac{\min_{\ell\in C}c_{j}}{1}
jAai3+(kBbi)/31+(Cci)/31\displaystyle\leq\frac{\sum_{j\in A}a_{i}}{3}+\frac{\big{(}\sum_{k\in B}b_{i}\big{)}/3}{1}+\frac{\big{(}\sum_{\ell\in C}c_{i}\big{)}/3}{1}
jAai+kBbi+Cci333=1\displaystyle\leq\frac{\sum_{j\in A}a_{i}+\sum_{k\in B}b_{i}+\sum_{\ell\in C}c_{i}}{3}\leq\frac{3}{3}=1

in the above, the second inequality follows since the minimum value is upper bounded by the average. The last inequality follows since Δu=1\Delta_{u}=1 and therefore the expected number of probes any offline vertex receives cannot exceed 1 and we have |U|=3|U|=3 many vertices.

To prove the same result for individual fairness we use the same graph. We note that the arrival of vertices in VV is KAD instead with the ithi^{\text{th}} vertex viv_{i} having pvi,i=1p_{v_{i},i}=1 and pvi,t=0,tip_{v_{i},t}=0,\forall t\neq i. Then we follow an argument similar to the above. ∎

The following proves Theorem 4.5 therefore showing that there is indeed a conflict between achieving group and individual fairness even if we were to consider only one side of the graph.

See 4.5

Proof.

Let us focus on the offline side, i.e. we consider αG\mathnormal{\alpha_{G}} and αI\mathnormal{\alpha_{I}} that are the competitive ratios for the group and individual fairness of the offline side.

Consider a graph which consists of two offline vertices and one online vertex, i.e. |U|=2|U|=2 and |V|=1|V|=1. Further, there is only one group. Let pe=1,eEp_{e}=1,\forall e\in E and uU,vV:Δu=Δv=1\forall u\in U,\forall v\in V:\mathnormal{\Delta_{\mathnormal{u}}}=\mathnormal{\Delta_{\mathnormal{v}}}=1. UU has two vertices u1u_{1} and u2u_{2} both connected to the same vertex vVv\in V. For edge (u1,v)(u_{1},v), we let w(u1,v)U=1w^{U}_{(u_{1},v)}=1 and for edge (u2,v)(u_{2},v), we let w(u2,v)U=Lw^{U}_{(u_{2},v)}=L where LL is an arbitrarily large number. Note that both of these weights are for the utility of the offline side. Finally, we only have one round so T=1T=1.

Let θ1\theta_{1} and θ2\theta_{2} be the expected number of probes edges (u1,v)(u_{1},v) and (u2,v)(u_{2},v) receive, respectively. Note that θ1=1θ2\theta_{1}=1-\theta_{2}. It follows that the optimal offline group fairness objective is OPTGU=maxθ1,θ2(θ1+Lθ2)=maxθ2((1θ2)+Lθ2)=L\operatorname{\mathrm{OPT}}^{U}_{G}=\max\limits_{\theta_{1},\theta_{2}}(\theta_{1}+L\theta_{2})=\max\limits_{\theta_{2}}((1-\theta_{2})+L\theta_{2})=L. Further, the optimal offline individual fairness objective is OPTIU=min{θ1,Lθ2}\operatorname{\mathrm{OPT}}^{U}_{I}=\min\{\theta_{1},L\theta_{2}\}, it is not difficult to show that OPTIU=LL+1\operatorname{\mathrm{OPT}}^{U}_{I}=\frac{L}{L+1}. Now consider the sum of competitive ratios, we have:

ALGGUOPTGU+ALGIUOPTIU\displaystyle\frac{\operatorname{\mathrm{ALG}}^{U}_{G}}{\operatorname{\mathrm{OPT}}^{U}_{G}}+\frac{\operatorname{\mathrm{ALG}}^{U}_{I}}{\operatorname{\mathrm{OPT}}^{U}_{I}} =θ1+Lθ2L+min{θ1,Lθ2}LL+1\displaystyle=\frac{\theta_{1}+L\theta_{2}}{L}+\frac{\min\{\theta_{1},L\theta_{2}\}}{\frac{L}{L+1}}
θ1+Lθ2L+θ1(L+1)L\displaystyle\leq\frac{\theta_{1}+L\theta_{2}}{L}+\frac{\theta_{1}(L+1)}{L}
=(L+2)θ1+Lθ2L\displaystyle=\frac{(L+2)\theta_{1}+L\theta_{2}}{L}
=(θ1+θ2)+2θ1L\displaystyle=(\theta_{1}+\theta_{2})+\frac{2\theta_{1}}{L}
1+2θ1LL1\displaystyle\leq 1+\frac{2\theta_{1}}{L}\xrightarrow{L\rightarrow\infty}1

this proves the result for the offline side of the graph.

To prove the result for the online side, we reverse the graph construction, i.e. having one vertex in UU and two vertex types in VV which arrive with equal probability. It now holds that OPTIV=min{θ1,Lθ2}\operatorname{\mathrm{OPT}}^{V}_{I}=\min\{\theta_{1},L\theta_{2}\} and by setting TT to an arbitrarily large value OPTGV=L\operatorname{\mathrm{OPT}}^{V}_{G}=L. Then we follow an identical argument to the above. ∎

Appendix C Additonal Experimental Results

As mentioned before one of the major contributions of our work is that we consider the operator’s profit and fairness for both sides simultaneously instead of fairness for only one side. To further see the effects of ignoring one side, we run TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} with one side ignored (see table 2). It is clear that the fairness objective for the ignored side is indeed lower in comparison to what can be achieved in figure 1. More precisely, we can see that the Offline (Driver) and Online (Rider) fairness can be simultaneously improved to around 0.50.5 by setting α=0.5,β=γ=0.25\alpha=0.5,\beta=\gamma=0.25 (figure 1) whereas their values when their optimization weight is set to zero is 0.3870.387 and 0.410.41 , respectively (see table 2).

Profit Driver Fairness Rider Fairness
α=γ=0.5,β=0\alpha=\gamma=0.5,\beta=0 0.43 0.387 0.509
α=β=0.5,γ=0\alpha=\beta=0.5,\gamma=0 0.564 0.498 0.41
Table 2: Results of running TSGFKIID\operatorname{\mathrm{TSGF}_{\textbf{KIID}}} on the NYC dataset with the fairness on one side ignored, i.e. its optimization weight set to 0: (Top Row) Offline (Driver) fairness ignored (α=γ=0.5,β=0\alpha=\gamma=0.5,\beta=0) and (Bottom Row) Online (Rider) fairness ignored (α=0.5,β=0.5,γ=0\alpha=0.5,\beta=0.5,\gamma=0).