This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On Bounds for Greedy Schemes in String Optimization based on Greedy Curvatures

Bowen Li, Brandon Van Over, Edwin K. P. Chong, and Ali Pezeshki This work is supported in part by the AFOSR under award FA8750-20- 2-0504 and by the NSF under award CNS-2229469.B. Li, B. Van Over, E. K. P. Chong and A. Pezeshki are with the Department of Electrical and Computer Engineering, Colorado State University, Fort Collins, CO 80523, USA. {bowen.li, b.van_over, edwin.chong, ali.pezeshki}@colostate.edu
Abstract

We consider the celebrated bound introduced by Conforti and Cornuéjols (1984) for greedy schemes in submodular optimization. The bound assumes a submodular function defined on a collection of sets forming a matroid and is based on greedy curvature. We show that the bound holds for a very general class of string problems that includes maximizing submodular functions over set matroids as a special case. We also derive a bound that is computable in the sense that they depend only on quantities along the greedy trajectory. We prove that our bound is superior to the greedy curvature bound of Conforti and Cornuéjols. In addition, our bound holds under a condition that is weaker than submodularity.

I INTRODUCTION

In many sequential decision-making or machine learning problems, we encounter the problem of optimally choosing a string (ordered set) of actions over a finite horizon to maximize a given objective function. String optimization problems have the added complexity relative to set optimization problems in that the objective function depends on both the set of actions taken and the order in which the actions are taken. In such problems, determining the optimal solution (optimal string of actions) can become computationally intractable with increasing size of state/action space and optimization horizon. Therefore, we often have to resort to approximate solutions. One of the most common approximation schemes is the greedy scheme, in which we sequentially select the action that maximizes the increment in the objective function at each step. A natural question that arises is how good is the greedy solution to a string optimization problem relative to the optimal solution?

In this paper, we derive a ratio bound for the performance of greedy solutions to string optimization problems relative to that of the optional solution. By a ratio bound, we mean a bound of the form

f(GK)f(OK)β,\frac{f(G_{K})}{f(O_{K})}\geq\beta,

where f(GK)f(G_{K}) and f(OK)f(O_{K}) are the objective function values of the greedy solution and optimal solution, respectively, and KK is the horizon. The bound guarantees that the greedy solution achieves at least a factor of β\beta of the optimal solution. Our focus is on establishing a bound in which the factor β\beta is easily computable.

Prior work: Bounding greedy solutions to set optimization problems has a rich history, especially in the context of submodular set functions. The most celebrated results are from Fisher et al.  [1] and Nemhauser et al. [2]. They showed that β1/2\beta\geq 1/2 over a finite general set matroid [1] and β1((K1)/K)K>(1e1)\beta\geq 1-((K-1)/K)^{K}>(1-e^{-1}) over a finite uniform set matroid with KK being the horizon [2]. Improved bounds (with β\beta larger than 1e11-e^{-1}) have been developed by Conforti and Cornuéjols [3], Vondrak [4, 5], and Wang et al. [6], by introducing various notions of curvature for set submodular functions under uniform and/or general matroid settings.

However, for most of these bounds, computing the value of β\beta is intractable and hence the bound is effectively not computable in problems with large action spaces and decision horizons. In addition, the value of β\beta depends on the values of the objective function beyond the optimization horizon. One notable exception is the bound derived by Conforti and Cornuéjols [3] that is based on a quantity called greedy curvature. This bound is easily computable. Another notable exception is the bound derived by Welikala et al. [7], which is also computable and in some problems has shown to be larger than other curvature bounds. However, the value of β\beta for this bound depends on values of the objective function beyond the optimization horizon.

Recently, Zhang et al. [8] extended the concept of set submodularity to string submodularity and established a β=(1e1)\beta=(1-e^{-1}) bound for greedy solutions of string optimization problems. They also derived improved bounds involving various notions of curvature for string submodular functions. However, these curvatures have the same computational intractability issue as mentioned above. Alaei et al. [9] proved (1e1)(1-e^{-1}) and (11/e(11/e))(1-1/e^{(1-1/e)}) bounds for online advertising and query rewriting problems, both of which can be formulated under the framework of string submodular optimization.

Main contributions: The contributions of this paper are as follows:

  1. 1.

    We first derive a computable bound for greedy solutions to string optimization problems by extending the notion of greedy curvature and the proof technique used by Conforti and Cornuéjols [3] from set functions to string functions. Our bound coincides with the greedy curvature bound in [3] if reducing our string objective function to a set objective function. However, in deriving our bound, we don’t use submodularity. Rather we rely on weaker conditions on the string objective function. Therefore, our bounding result is stronger. This result is reported in \threfGenConforti.

  2. 2.

    We then establish an even stronger result. We derive another computable bound that relies on weaker assumptions than those used in deriving our first bound. This is done in \threftopK.

  3. 3.

    We then show that the value β2\beta_{2} of the bound in \threftopK is larger than the value β1\beta_{1} of the bound in \threfGenConforti (i.e., the generalization of the Conforti and Cornuéjols bound). This is shown in \threfbetterThm.

Organization: The paper is organized as follows. Section II introduces all the mathematical preliminaries and formulates the string optimization problem. Section III presents the mains results regarding our computable bounds. Applications of our theoretical results to task scheduling and sensor coverage problems are demonstrated in Section IV. Finally, conclusions are given in Section V.

II DEFINITIONS FOR STRING OPTIMIZATION

In this section, we introduce notation, terminology, and definitions that we use in our paper and present a formulation for a general string optimization problem.

Definition 1.

Let 𝕊\mathbb{S} be a set. In our context, 𝕊\mathbb{S} is called the ground set and its elements are called symbols.

  1. 1.

    Let s1,s2,,sk𝕊s_{1},s_{2},\ldots,s_{k}\in\mathbb{S}. Then S:=s1s2skS:=s_{1}s_{2}\cdots s_{k} is a string with length |S|=k|S|=k.

  2. 2.

    Let KK be a positive integer, called the horizon.

  3. 3.

    Let 𝕊K\mathbb{S}_{K} be the set of all positive strings of length up to KK, including the empty string \varnothing. It is also called the uniform string matroid of rank KK.

Definition 2.

Consider two strings S=s1s2smS=s_{1}s_{2}\cdots s_{m} and T=t1t2tnT=t_{1}t_{2}\cdots t_{n} in 𝕊K\mathbb{S}_{K}.

  1. 1.

    Define the concatenation of SS and TT as ST:=s1s2smt1t2tnST:=s_{1}s_{2}\cdots s_{m}t_{1}t_{2}\cdots t_{n}

  2. 2.

    We say that PP is a prefix of SS if S=PUS=PU for some U𝕊KU\in\mathbb{S}_{K}, in which we write PSP\preceq S.

  3. 3.

    We say that 𝕋𝕊K\mathbb{T}\subseteq\mathbb{S}_{K} is prefix-closed if for all S𝕋S\in\mathbb{T} and PSP\preceq S, P𝕋P\in\mathbb{T}.

  4. 4.

    Let f:𝕊Kf:\mathbb{S}_{K}\rightarrow\mathbb{R} be the objective function.

Definition 3.

Let 𝕏𝕊K\mathbb{X}\subseteq\mathbb{S}_{K}. Then (𝕏,𝕊)(\mathbb{X},\mathbb{S}) is a finite rank KK string matroid if

  1. 1.

    |A|K|A|\leq K for all A𝕏A\in\mathbb{X}.

  2. 2.

    If B𝕏B\in\mathbb{X} and ABA\preceq B, then A𝕏A\in\mathbb{X}.

  3. 3.

    For every A,B𝕏A,B\in\mathbb{X} where |A|+1=|B||A|+1=|B|, there exists bb that is a symbol in BB and Ab𝕏Ab\in\mathbb{X}.

Definition 4.

f:𝕊Kf:\mathbb{S}_{K}\rightarrow\mathbb{R} is string submodular on (𝕏,𝕊)(\mathbb{X},\mathbb{S}) if

  1. 1.

    AB𝕏\forall A\preceq B\in\mathbb{X}, f(A)f(B)f(A)\leq f(B).

  2. 2.

    AB𝕏 and j𝕊\forall A\preceq B\in\mathbb{X}\text{ and }\forall j\in\mathbb{S} such that Aj,Bj𝕏Aj,Bj\in\mathbb{X}, f(Aj)f(A)f(Bj)f(B)f(Aj)-f(A)\geq f(Bj)-f(B).

Remark 1.
  1. 1.

    Unlike the permutation invariance in the set case, the order of the symbols in a string matters. Different orders of the same set of symbols represent different strings.

  2. 2.

    Note that ST𝕊KST\in\mathbb{S}_{K} if and only if m+nKm+n\leq K.

  3. 3.

    In the matroid literature, a prefix-closed collection is also said to satisfy the hereditary property and is called as independence system. An independence system does not require condition (3)(3) in Definition 3.

Remark 2.

The definitions of finite rank string matroid and string submodular function are introduced in [10]. Their formulations are inspired by the definitions of set matroid and set submodular function in [3]. Most of the previous work is established in the set case. Here, we extend the theoretical results to the more general string case.

Consider an objective function ff and a prefix-closed 𝕋\mathbb{T} with maxS𝕋|S|=K\max_{S\in\mathbb{T}}|S|=K. The general string-optimization is given by

maximize f(S)\displaystyle\text{maximize }f(S) (1)
subject to S𝕋\displaystyle\text{subject to }S\in\mathbb{T}
Remark 3.
  1. 1.

    The stipulation that f()=0f(\varnothing)=0 is without loss of generality because if f()0f(\varnothing)\neq 0, we can subtract f()f(\varnothing) from all values of ff without changing the maximizer of the optimization problem (1).

  2. 2.

    We deal with the case that f(S)0f(S)\geq 0 for any S𝕋S\in\mathbb{T}.

  3. 3.

    The constraint set 𝕋\mathbb{T} here is not necessarily the uniform matroid 𝕊K\mathbb{S}_{K} or a finite rank KK string matroid (𝕏,𝕊)(\mathbb{X},\mathbb{S}). In particular, our analysis applies to general prefix-closed sets as constraint sets subject to certain assumptions.

  4. 4.

    The stipulation that maxS𝕋|S|=K\max_{S\in\mathbb{T}}|S|=K is without loss of generality because we can always define K:=maxS𝕋|S|K:=\max_{S\in\mathbb{T}}|S|.

Definition 5.

Any solution to the optimization problem (1) is said to be optimal, denoted by OL=o1o2oLO_{L}=o_{1}o_{2}\cdots o_{L}, where L{1,,K}L\in\{1,\ldots,K\}.

Remark 4.
  1. 1.

    The length LL of an optimal solution OLO_{L} could be anything from 11 to KK.

  2. 2.

    There may be multiple optimal solutions.

  3. 3.

    If 𝕊\mathbb{S} is finite, then 𝕊K\mathbb{S}_{K} and 𝕋\mathbb{T} are finite, and an optimal solution always exists.

Definition 6.

We define GK=g1g2gKG_{K}=g_{1}g_{2}\cdots g_{K} to be a greedy solution if for all k{1,2,,K}k\in\{1,2,\ldots,K\},

gk=argmaxs𝕊:g1gk1s𝕋f(g1gk1s).g_{k}=\operatorname*{arg\,max}_{s\in\mathbb{S}:g_{1}\cdots g_{k-1}s\in\mathbb{T}}f(g_{1}\cdots g_{k-1}s).

Note that we are going all the way to horizon KK. This implicitly assumes that ff is nondecreasing along GKG_{K} and f(g1)>0f(g_{1})>0. Theses implicit conditions are reflected in assumption 𝐀𝟑\mathbf{A_{3}} in Section III.

Remark 5.
  1. 1.

    The length of a greedy solution GKG_{K} is always KK.

  2. 2.

    A greedy scheme is the one that adds a symbol to the existing string at each time k for k{1,,K1}k\text{ for }k\in\{1,\ldots,K-1\}, so that the resulting string produces the highest value of ff without regard to the future times. In other words, a greedy solution is generated by a greedy scheme.

  3. 3.

    The argmax\operatorname*{arg\,max} above could be nonunique, in which case there would be multiple greedy solutions.

Definition 7.

We define β:=f(GK)/f(OK)\beta:=f(G_{K})/f(O_{K}) as the performance bound of the greedy solution.

Remark 6.

f(GK)f(G_{K}) can be computed exactly, but f(OK)f(O_{K}) is computational intractable. Finding a valid lower bound of β\beta is equivalent to finding an upper bound of f(OK)f(O_{K}).

III Main Results

In this section, we present our bounding results. We first introduce the assumptions that our bounding results rely on. Then we generalize the greedy curvature bound of Conforti and Cornuéjols [3] from set optimization problems to string optimization problems. Then, we establish a stronger bounding result that relies on fewer assumptions and has a larger bound value.

Given any string S=s1s2s|S|𝕊KS=s_{1}s_{2}\cdots s_{|S|}\in\mathbb{S}_{K} and i,j{1,2,,|S|}i,j\in\{1,2,\ldots,|S|\}, denote Si:j:=sisjS_{i:j}:=s_{i}\cdots s_{j} if iji\leq j and Si:j=S_{i:j}=\varnothing if i>ji>j. For simplicity, we use the abbreviation Sk:=S1:k(k{0,,|S|})S_{k}:=S_{1:k}\;(k\in\{0,\ldots,|S|\}). Note that by definition, S0=S_{0}=\varnothing.

Let S𝕋S\in\mathbb{T}. Define the Δ\Delta notation as follows: for each k{1,,|S|}k\in\{1,\ldots,|S|\}, Δ(Sk):=f(Sk)f(Sk1)\Delta(S_{k}):=f(S_{k})-f(S_{k-1}), called the kthk^{\text{th}} increment of f(S)f(S). In [3], Δ(Sk)\Delta(S_{k}) is called a discrete derivative.

For any S𝕋S\in\mathbb{T}, let 𝕊(S):={s𝕊:Ss𝕋}\mathbb{S}(S):=\{s\in\mathbb{S}:Ss\in\mathbb{T}\}. The set 𝕊(S)\mathbb{S}(S) contains the actions that are feasible with respect to 𝕋\mathbb{T} following string SS. The set 𝕊(Gk1)={s𝕊:Gk1s𝕋}\mathbb{S}(G_{k-1})=\{s\in\mathbb{S}:G_{k-1}s\in\mathbb{T}\} is frequently referred to in our derivations. When k=1k=1, 𝕊(Gk1)=𝕊()\mathbb{S}(G_{k-1})=\mathbb{S}(\varnothing). For k>1k>1, 𝕊(Gk1)={s𝕊:Gk1s𝕋}\mathbb{S}(G_{k-1})=\{s\in\mathbb{S}:G_{k-1}s\in\mathbb{T}\} is nonemepty because gk𝕊(Gk1)g_{k}\in\mathbb{S}(G_{k-1}) by definition.

Assumptions: We introduce the following three key assumptions regarding feasibility along the greedy sequence and the diminishing return property along the optimal and greedy paths:

𝐀𝟏:\displaystyle\mathbf{A_{1}:} For each k{1,,K},ok𝕊(Gk1).\displaystyle\text{ For each }k\in\{1,\ldots,K\},o_{k}\in\mathbb{S}(G_{k-1}).
𝐀𝟐:\displaystyle\mathbf{A_{2}:} For each k{1,,K},Δ(Ok)f(ok).\displaystyle\text{ For each }k\in\{1,\ldots,K\},\Delta(O_{k})\leq f(o_{k}).
𝐀𝟑:\displaystyle\mathbf{A_{3}:} For each k{1,,K},Δ(Gk1s)>0 for all\displaystyle\text{ For each }k\in\{1,\ldots,K\},\Delta(G_{k-1}s)>0\text{ for all}
s𝕊(Gk1).\displaystyle\;s\in\mathbb{S}(G_{k-1}).

Our bounding results (Theorem 1-3) rely on a subset or all of the the above three assumptions. We note that 𝐀𝟐\mathbf{A_{2}} is a weaker conditions than string submodularity, they only involve diminishing return along the optimal sequence instead of any sequence in 𝕋\mathbb{T}. Any string submodular function satisfies 𝐀𝟐\mathbf{A_{2}}, as stated in the following lemma.

Lemma 1.

If ff is string submodular, then ff satisfies 𝐀𝟐\mathbf{A_{2}}.

Proof.

By condition (2) of string submodularity in Definition 4, we can set A= and B=Ok1 for k{1,,K}A=\varnothing\text{ and }B=O_{k-1}\text{ for }k\in\{1,\ldots,K\}. Then Aok=okAo_{k}=o_{k} and Bok=OkBo_{k}=O_{k} are both feasible in the domain of ff and f(Ok)f(Ok1)=Δ(Ok)f(ok) for k{1,,K}f(O_{k})-f(O_{k-1})=\Delta(O_{k})\leq f(o_{k})\text{ for }k\in\{1,\ldots,K\}. ∎

We now establish our first bounding result. We define the greedy curvature for a string function as

αG\displaystyle\alpha_{G} :=maxk{2,,K}maxs{s|Δ(Gk1s)>0,s𝕊(Gk1)}f(s)Δ(Gk1s).\displaystyle:=\max_{k\in\{2,\ldots,K\}}\max_{s\in\{s|\Delta(G_{k-1}s)>0,s\in\mathbb{S}(G_{k-1})\}}\frac{f(s)}{\Delta(G_{k-1}s)}. (2)
Remark 7.
\thlabel

alpha_G_1 Based on assumptions 𝐀𝟐\mathbf{A_{2}} and 𝐀𝟑\mathbf{A_{3}}, αG1\alpha_{G}\geq 1 always holds. This can be proved by contradiction. Assume αG<1\alpha_{G}<1. By the definition of αG\alpha_{G} and assumptions 𝐀𝟐\mathbf{A_{2}} and 𝐀𝟑\mathbf{A_{3}}, we have:

αGΔ(Gk1gk)=αGΔ(Gk)f(gk)f(ok)Δ(OK)\displaystyle\alpha_{G}\Delta(G_{k-1}g_{k})=\alpha_{G}\Delta(G_{k})\geq f(g_{k})\geq f(o_{k})\geq\Delta(O_{K}) (3)
for all k{2,,K}.\displaystyle\text{for all }k\in\{2,\ldots,K\}.

Adding f(o1)f(g1)f(o_{1})\leq f(g_{1}) and summing over αGΔ(Gk)Δ(OK)\alpha_{G}\Delta(G_{k})\geq\Delta(O_{K}) for all k{2,,K}k\in\{2,\ldots,K\} yield:

f(OK)\displaystyle f(O_{K}) =f(o1)+k=2KΔ(OK)f(g1)+k=2KαGΔ(Gk)\displaystyle=f(o_{1})+\sum_{k=2}^{K}\Delta(O_{K})\leq f(g_{1})+\sum_{k=2}^{K}\alpha_{G}\Delta(G_{k}) (4)
<f(g1)+k=2KΔ(Gk)=f(GK),\displaystyle<f(g_{1})+\sum_{k=2}^{K}\Delta(G_{k})=f(G_{K}),

which brings contradiction with f(OK)<f(GK)f(O_{K})<f(G_{K}).

Then, we have the following theorem.

Theorem 1.
\thlabel

GenConforti Assuming 𝐀𝟏,𝐀𝟐 and 𝐀𝟑\mathbf{A_{1}},\mathbf{A_{2}}\text{ and }\mathbf{A_{3}}, f(GK)/f(OK)β1f(G_{K})/f(O_{K})\geq\beta_{1}, where

β1=1K+1αGK1K.\beta_{1}=\frac{1}{K}+\frac{1}{\alpha_{G}}*\frac{K-1}{K}.
Proof.

The definition of αG\alpha_{G} gives us:

αGf(s)Δ(Gk1s)\displaystyle\alpha_{G}\geq\frac{f(s)}{\Delta(G_{k-1}s)}

for every k{2,,K}k\in\{2,\ldots,K\} and s{s|Δ(Gk1s)>0,s𝕊(Gk1)}.s\in\{s|\Delta(G_{k-1}s)>0,s\in\mathbb{S}(G_{k-1})\}.

By assumption 𝐀𝟏\mathbf{A_{1}},

f(ok)αGΔ(Gk1ok) for k{2,,K}.\displaystyle f(o_{k})\leq\alpha_{G}\Delta(G_{k-1}o_{k})\text{ for }k\in\{2,\ldots,K\}. (5)

Moreover, by assumption 𝐀𝟐\mathbf{A_{2}},

f(OK)=k=1KΔ(Ok)k=1Kf(ok)(a)Kf(g1),f(O_{K})=\sum_{k=1}^{K}\Delta(O_{k})\leq\sum_{k=1}^{K}f(o_{k})\stackrel{{\scriptstyle(a)}}{{\leq}}Kf(g_{1}), (6)

in which (a)(a) holds since f(g1)f(g_{1}), by the definition of greedy solution, dominates each term in k=1Kf(ok)\sum_{k=1}^{K}f(o_{k}).

Then,

f(OK)\displaystyle f(O_{K}) =k=1K(f(Ok)f(Ok1))=k=1KΔ(Ok)\displaystyle=\sum_{k=1}^{K}(f(O_{k})-f(O_{k-1}))=\sum_{k=1}^{K}\Delta(O_{k}) (7)
k=1Kf(ok)(b)f(g1)+k=2KαGΔ(Gk1ok)\displaystyle\leq\sum_{k=1}^{K}f(o_{k})\stackrel{{\scriptstyle(b)}}{{\leq}}f(g_{1})+\sum_{k=2}^{K}\alpha_{G}\Delta(G_{k-1}o_{k})
(c)f(g1)+k=2KαGΔ(Gk)\displaystyle\stackrel{{\scriptstyle(c)}}{{\leq}}f(g_{1})+\sum_{k=2}^{K}\alpha_{G}\Delta(G_{k})
=αG(f(g1)+k=2KΔ(GK))+(1αG)f(g1)\displaystyle=\alpha_{G}(f(g_{1})+\sum_{k=2}^{K}\Delta(G_{K}))+(1-\alpha_{G})f(g_{1})
=αGf(GK)+(1αG)f(g1),\displaystyle=\alpha_{G}f(G_{K})+(1-\alpha_{G})f(g_{1}),

where (b)(b) is by the relation in inequality (5)(\ref{beta_greedy_1}), and (c)(c) is by the definition of gkg_{k}.

Dividing αGf(OK)\alpha_{G}f(O_{K}) on both sides of (7)(\ref{beta_greedy_3}) yields:

f(GK)f(OK)\displaystyle\frac{f(G_{K})}{f(O_{K})} 1αG1αGαGf(g1)f(OK)(d)1αG1αGαG1K\displaystyle\geq\frac{1}{\alpha_{G}}-\frac{1-\alpha_{G}}{\alpha_{G}}*\frac{f(g_{1})}{f(O_{K})}\stackrel{{\scriptstyle(d)}}{{\geq}}\frac{1}{\alpha_{G}}-\frac{1-\alpha_{G}}{\alpha_{G}}*\frac{1}{K} (8)
f(GK)f(OK)\displaystyle\frac{f(G_{K})}{f(O_{K})} 1K+1αGK1K,\displaystyle\geq\frac{1}{K}+\frac{1}{\alpha_{G}}*\frac{K-1}{K},

where (d)(d) is due to inequality (6)(\ref{beta_greedy_2}) and the fact that αG1\alpha_{G}\geq 1 from \threfalpha_G_1. ∎

We note that the above proof technique is similar to the proof technique used in Theorem 3.1 in [3] by Conforti and Cornuéjols, in deriving their greedy curvature bound for set optimization problems. However, Theorem 1 is more general in that it applies to bounding greedy solutions in string optimization problems, which subsume set optimization problems, and it does not rely on having submodularity. The weaker assumptions 𝐀𝟐\mathbf{A_{2}} suffices. For a set submodular function, the value of the bound in Theorem 1 coincides with the value of the bound of Conforti and Cornuéjols. The result in Theorem 1 also generalizes our prior result for bounding string optimizations [10].

We now establish a stronger result.

Theorem 2.
\thlabel

topK Assuming 𝐀𝟏\mathbf{A_{1}} and 𝐀𝟐\mathbf{A_{2}}, f(GK)/f(OK)β2f(G_{K})/f(O_{K})\geq\beta_{2}, where

β2=f(GK)k=1Kmaxs𝕊(Gk1)f(s).\beta_{2}=\frac{f(G_{K})}{\sum_{k=1}^{K}\max_{s\in\mathbb{S}(G_{k-1})}f(s)}. (9)
Proof.
f(OK)=k=1K(f(Ok)f(Ok1))=k=1KΔ(Ok)\displaystyle f(O_{K})=\sum_{k=1}^{K}(f(O_{k})-f(O_{k-1}))=\sum_{k=1}^{K}\Delta(O_{k})

By 𝐀𝟐\mathbf{A_{2}}, Δ(Ok)f(ok)\Delta(O_{k})\leq f(o_{k}) for k=1,,Kk=1,\ldots,K. Therefore,

k=1KΔ(Ok)(a)k=1Kf(ok)(b)k=1Kmaxs𝕊(Gk1)f(s),\sum_{k=1}^{K}\Delta(O_{k})\stackrel{{\scriptstyle(a)}}{{\leq}}\sum_{k=1}^{K}f(o_{k})\stackrel{{\scriptstyle(b)}}{{\leq}}\sum_{k=1}^{K}\max_{s\in\mathbb{S}(G_{k-1})}f(s),

where (a)(a) is by 𝐀𝟐\mathbf{A_{2}} and (b)(b) is by 𝐀𝟏\mathbf{A_{1}} and the definition of the max\max operator. ∎

Theorem 2 essentially shows that the sum of the objective values under the first KK greedy and feasible actions serves a valid upper bound for f(OK)f(O_{K}). This upper bound is easy to compute along the greedy trajectory and is obtained under weaker assumptions than those in Theorem 1.

We now show that the bound in Theorem 2 is better than the bound in Theorem 1.

Theorem 3.
\thlabel

betterThm Assuming 𝐀𝟏,𝐀𝟐 and 𝐀𝟑\mathbf{A_{1}},\mathbf{A_{2}}\text{ and }\mathbf{A_{3}}, β2β1\beta_{2}\geq\beta_{1}.

Proof.

Let sks_{k} denote the maximizer of f(s)f(s) for s𝕊(Gk1)s\in\mathbb{S}(G_{k-1}) where k{1,,K}k\in\{1,\ldots,K\}. Then k=1Kmaxs𝕊(Gk1)f(s)=k=1Kf(sk)Kf(g1)\sum_{k=1}^{K}\max_{s\in\mathbb{S}(G_{k-1})}f(s)=\sum_{k=1}^{K}f(s_{k})\leq Kf(g_{1}) by the definition of g1g_{1}. The definition of αG\alpha_{G} further gives us:

k=1Kf(sk)\displaystyle\sum_{k=1}^{K}f(s_{k}) f(g1)+k=2KαGΔ(Gk1sk)\displaystyle\leq f(g_{1})+\sum_{k=2}^{K}\alpha_{G}\Delta(G_{k-1}s_{k}) (10)
f(g1)+k=2KαGΔ(Gk)\displaystyle\leq f(g_{1})+\sum_{k=2}^{K}\alpha_{G}\Delta(G_{k})
=αGf(GK)+(1αG)f(g1).\displaystyle=\alpha_{G}f(G_{K})+(1-\alpha_{G})f(g_{1}).

Dividing αGk=1Kf(sk)\alpha_{G}\sum_{k=1}^{K}f(s_{k}) on both sides of inequality (10) yields:

β2=f(GK)k=1Kf(sk)\displaystyle\beta_{2}=\frac{f(G_{K})}{\sum_{k=1}^{K}f(s_{k})} 1αG1αGαGf(g1)k=1Kf(sk)\displaystyle\geq\frac{1}{\alpha_{G}}-\frac{1-\alpha_{G}}{\alpha_{G}}*\frac{f(g_{1})}{\sum_{k=1}^{K}f(s_{k})} (11)
(a)1αG1αGαG1K\displaystyle\stackrel{{\scriptstyle(a)}}{{\geq}}\frac{1}{\alpha_{G}}-\frac{1-\alpha_{G}}{\alpha_{G}}*\frac{1}{K}
=1K+1αGK1K=β1,\displaystyle=\frac{1}{K}+\frac{1}{\alpha_{G}}*\frac{K-1}{K}=\beta_{1},

where inequality (a) is due to the fact that k=1Kf(sk)Kf(g1)\sum_{k=1}^{K}f(s_{k})\leq Kf(g_{1}) and αG1\alpha_{G}\geq 1 from \threfalpha_G_1. ∎

Remark 8.

Recall the improved greedy curvature αk\alpha_{k} proposed in [10]. For k{2,,K}k\in\{2,\ldots,K\},

αk:=maxs{s|Δ(Gk1s)>0,s𝕊(Gk1)}f(s)Δ(Gk1s).\alpha_{k}:=\max_{s\in\{s|\Delta(G_{k-1}s)>0,s\in\mathbb{S}(G_{k-1})\}}\frac{f(s)}{\Delta(G_{k-1}s)}.

By the definition of αk\alpha_{k}, f(sk)αkΔ(Gk1sk)αkΔ(Gk)f(s_{k})\leq\alpha_{k}\Delta(G_{k-1}s_{k})\leq\alpha_{k}\Delta(G_{k}) for k{2,,K}k\in\{2,\ldots,K\}. Hence,

k=1Kf(sk)f(g1)+k=2KαkΔ(Gk),\sum_{k=1}^{K}f(s_{k})\leq f(g_{1})+\sum_{k=2}^{K}\alpha_{k}\Delta(G_{k}), (12)

The right hand side of (12) is the upper bound of f(OK)f(O_{K}) in [10], which is proven to give a better performance bound than β1\beta_{1}. Therefore, β2\beta_{2} is also an improved performance bound than that proposed in [10].

IV APPLICATIONS

IV-A Task Scheduling

As a canonical optimization problem in operations research, task scheduling was studied in [11] and further investigated in [8] and [12]. In task scheduling, we aim to assign agents at multiple stages throughout the task to maximize the probability of successful completion. The detailed mathematical formulation is as follows.

Assume a task is comprised of KK stages, and an agent needs to be assigned at stage k(1kK)k\;(1\leq k\leq K) to accomplish the task. We have a pool of NN agents, denoted by M1,,MNM_{1},\ldots,M_{N}, and no agent is allowed to be repeatedly selected. For each agent Mi(1iN)M_{i}\;(1\leq i\leq N), the probability of accomplishing the task at each stage k(1kK)k\;(1\leq k\leq K) is denoted by p(mik)p(m_{i}^{k}). Therefore, the probability of accomplishment after selecting Mi1,Mi2,,MiKM_{i_{1}},M_{i_{2}},\ldots,M_{i_{K}} in the end is:

f(Mi1Mi2MiK):=1k=1K(1p(mikk)),f(M_{i_{1}}M_{i_{2}}\cdots M_{i_{K}}):=1-\prod_{k=1}^{K}(1-p(m_{i_{k}}^{k})), (13)

where the subscript ii_{*} in MM and mm represents the selected agent number. Note that the probability of accomplishing the task for each agent is stage-variant. f(Mi1Mi2MiK)f(M_{i_{1}}M_{i_{2}}\cdots M_{i_{K}}) in (13)(\ref{task_scheduling}) is henceforth a string function and order dependent. Streeter and Golovin proved in [11] that the greedy schemes yields a β0=(1e1)\beta_{0}=(1-e^{-1}) performance bound when ff bears submodularity and monotonicity.

Consider the following example with K=3K=3, N=5N=5, and the probabilities of successful completion in Table. I. The optimal sequence (O3)(O_{3}) and greedy sequence (G3)(G_{3}) are equal in this example with G3=O3=M1M2M3G_{3}=O_{3}=M_{1}M_{2}M_{3}. We can verify that assumption 𝐀𝟏\mathbf{A_{1}} is satisfied, and ff is string submodular. The three performance bounds are as follows:

β2=f(M1M2M3)f(M1)+f(M2)+f(M3)=0.7816;\displaystyle\beta_{2}=\frac{f(M_{1}M_{2}M_{3})}{f(M_{1})+f(M_{2})+f(M_{3})}=0.7816;
β1=1K+1αGK1K=0.5893;\displaystyle\beta_{1}=\frac{1}{K}+\frac{1}{\alpha_{G}}*\frac{K-1}{K}=0.5893;
β0=1e1=0.6321.\displaystyle\beta_{0}=1-e^{-1}=0.6321.

We observe that the β2\beta_{2} bound outperforms the other two bounds. This example shows the curvature bounds are not necessarily better than the original (1e1)(1-e^{-1}) bound. On some problems they are better, in others they are not. Therefore, we need to choose the largest of the three bounds. But this is easy, because all three bounds are easily compuatble. In the next example, we show a plot that compares the three bounds under different parameters of another optimization problem.

Stage 1 Stage 2 Stage 3
M1M_{1} 0.2 0.16 0.14
M2M_{2} 0.18 0.16 0.14
M3M_{3} 0.16 0.14 0.14
M4M_{4} 0.14 0.12 0.10
M5M_{5} 0.12 0.1 0.08
TABLE I: Probability of successful completion for each agent at each stage.

IV-B Multi-agent Sensor Coverage

The multi-agent sensor coverage problem was originally studied in [13] and further analyzed in [14] and [7]. In a given mission space, we need to find a placement of a set of homogeneous sensors to maximize the probability of detecting randomly occurring events. This problem can be formulated as a set optimization problem, which is a special case in the string setting. In this subsection, we demonstrate our theoretical results by applying them to a discrete version of the coverage problem. Our simplified version can be easily generalized to more complicated settings within the same theoretical framework.

The mission space Ω2\Omega\subset\mathbb{R}^{2} is modeled as a non-self-intersecting polygon where KK homogeneous sensors will be placed to detect a randomly occurring event in Ω\Omega. For simplicity of calculation, we assume both the sensors and the random event can only be placed and occur at lattice points. We denote the feasible space for sensor placement and event occurrence as ΩF\Omega^{F}. Our goal is to maximize the overall likelihood of successful detection in the mission space, as illustrated in Fig. 1.

Refer to caption
Figure 1: Sensor Coverage in a Mission Space

The likelihood of event occurrence over ΩF\Omega^{F} is given by an event mass function R:ΩF0R:\Omega^{F}\xrightarrow{}\mathbb{R}_{\geq 0}, and we assume that 𝐱ΩFR(𝐱)<\sum_{\mathbf{x}\in\Omega^{F}}R(\mathbf{x})<\infty. R(𝐱)R(\mathbf{x}) may reflect a particular distribution if some prior information is available. Otherwise, R(𝐱)=1R(\mathbf{x})=1 when no prior information is obtained. The locations of all the sensors are represented as 𝐬=(𝐬1,𝐬2,,𝐬K)(ΩF)K\mathbf{s}=(\mathbf{s}_{1},\mathbf{s}_{2},\ldots,\mathbf{s}_{K})\in(\Omega^{F})^{K}, where 𝐬i(1iK)\mathbf{s}_{i}\;(1\leq i\leq K) are coordinates of the placed sensors. Each sensor placed at 𝐬𝐢\mathbf{s_{i}} can detect any occurring event at location 𝐱\mathbf{x} with probability p(𝐱,𝐬i)=eλ𝐱𝐬ip(\mathbf{x},\mathbf{s}_{i})=e^{-\lambda\|\mathbf{x}-\mathbf{s}_{i}\|}, where λ\lambda is the decay rate characterizing how quick the sensing capability decays along the distance.

Assuming all the sensors are working independently, the probability of detecting an occurring event at location 𝐱ΩF\mathbf{x}\in\Omega^{F} after placing KK homogeneous sensors at locations 𝐬\mathbf{s} is p(𝐱,𝐬)=1i=1K(1p(𝐱,𝐬i))p(\mathbf{x},\mathbf{s})=1-\prod_{i=1}^{K}\left(1-p(\mathbf{x},\mathbf{s}_{i})\right). In order to consider the whole feasible space for event occurrence, we need to incorporate the event mass function. Our objective function becomes H(𝐬)=𝐱ΩFR(𝐱)p(𝐱,𝐬)H(\mathbf{s})=\sum_{\mathbf{x}\in\Omega^{F}}R(\mathbf{x})p(\mathbf{x},\mathbf{s}). The goal is to maximize H(𝐬)H(\mathbf{s}):

maximize H(𝐬)\displaystyle\text{maximize }H(\mathbf{s}) (14)
subject to 𝐬(ΩF)K.\displaystyle\text{subject to }\mathbf{s}\in(\Omega^{F})^{K}.

If nn lattice points in ΩF\Omega^{F} are feasible for sensor placement, we need to select KK out of nn locations with its complexity being n!/(K!(nK)!)n!/\left(K!(n-K)!\right). This becomes a set optimization problem, and exhaustive search is computationally intractable when nn is large. Therefore, greedy algorithm is an approach for an approximate solution in polynomial time. It was proved that the continuous version of H(𝐬)H(\mathbf{s}) is submodular in [14], and it is not difficult to verify that its discrete version is also submodular.

In our experiment, we consider a rectangular mission space Ω\Omega of size 50×4050\times 40. A set of KK homogeneous sensors are waiting to be deployed on those integer coordinates within Ω\Omega, denoted by ΩF\Omega^{F}. For a point 𝐩=(sx,sy)\mathbf{p}=(s_{x},s_{y}), the event mass function is given by R(𝐩)=(sx+sy)/(sx max+yy max)R(\mathbf{p})=\left(s_{x}+s_{y}\right)/\left(s_{x\text{ max}}+y_{y\text{ max}}\right), where sx max=50s_{x\text{ max}}=50 and sy max=40s_{y\text{ max}}=40 are the largest values of the xx and yy coordinates in the mission space. This event mass function implies that the random events are more likely to occur around the top right corner of the rectangular mission space.

A comparison of the performance bounds under different decay rate with K=5K=5 is shown in the upper figure of Fig. 2. Small decay rates imply good sensing capability and strong submodularity, under which the greedy scheme does not produce nearly optimal objective function value. In the upper figure of Fig. 2, the β2\beta_{2} bound (red line) always exceeds the β1\beta_{1} bound (blue line), illustrating \threfbetterThm. In addition, we can observe instances in which the β2\beta_{2} bound is larger than the β0=(1e1)\beta_{0}=(1-e^{-1}), while the β1\beta_{1} bound is below this value.

A comparison of performance bounds under different number of placed sensors with λ=1\lambda=1 is shown in the lower figure of Fig. 2. We can still observe that the β2\beta_{2} bound (red line) always exceeds the β1\beta_{1} bound (blue line) with significant advantages. Both β1\beta_{1} and β2\beta_{2} bound decrease as the number of placed sensors KK increases.

Refer to caption
Figure 2: Upper Figure: Performance bound comparison under different decay rates when the number of placed sensors K=5K=5; Lower Figure: Performance bound comparison under different number of placed sensors when the decay rate λ=1\lambda=1.

V CONCLUSION

We derived a computable bound β1\beta_{1} for greedy solutions to string optimization problems by extending the notion of greedy curvature and the proof technique used by Conforti and Cornuéjols [3] from set functions to string functions. However, in deriving our bound we did not use submodularity. Rather we relied on weaker conditions on the string objective function, and hence produced a stronger bounding result. We then derived another computable bound β2\beta_{2} that relies on even weaker assumptions than those used in deriving our first bound β1\beta_{1}, further strengthening the bounding result. We also showed that this second bound β2\beta_{2} has a larger value than the first bound β2\beta_{2}. We demonstrated the superiority of our bound on two applications, task scheduling and multi-agent sensor coverage.

References

  • [1] M. L. Fisher, G. L. Nemhauser, and L. A. Wolsey, An analysis of approximations for maximizing submodular set functions—II.   Springer Berlin Heidelberg, 1978.
  • [2] G. L. Nemhauser, L. A. Wolsey, and M. L. Fisher, “An analysis of approximations for maximizing submodular set functions—I,” Math. Program., vol. 14, pp. 265–294, 1978.
  • [3] M. Conforti and G. Cornuéjols, “Submodular set functions, matroids and the greedy algorithm: Tight worst-case bounds and some generalizations of the Rado-Edmonds theorem,” Discrete. Appl. Math., vol. 7, no. 3, pp. 251–274, 1984.
  • [4] J. Vondrák, “Submodularity and curvature: The optimal algorithm (combinatorial optimization and discrete algorithms),” RIMS Kokyuroku Bessatsu B, vol. 23, pp. 253–266, 2010.
  • [5] G. Calinescu, C. Chekuri, M. Pal, and J. Vondrák, “Maximizing a monotone submodular function subject to a matroid constraint,” SIAM J. Comput., vol. 40, no. 6, pp. 1740–1766, 2011.
  • [6] Z. Wang, B. Moran, X. Wang, and Q. Pan, “Approximation for maximizing monotone non-decreasing set functions with a greedy method,” J. Comb. Optim., vol. 31, pp. 29–43, 2016.
  • [7] S. Welikala, C. G. Cassandras, H. Lin, and P. J. Antsaklis, “A new performance bound for submodular maximization problems and its application to multi-agent optimal coverage problems,” Automatica, vol. 144, p. 110493, 2022.
  • [8] Z. Zhang, E. K. P. Chong, A. Pezeshki, and W. Moran, “String submodular functions with curvature constraints,” IEEE Trans. Autom. Control, vol. 61, no. 3, pp. 601–616, 2015.
  • [9] S. Alaei, A. Makhdoumi, and A. Malekian, “Maximizing sequence-submodular functions and its application to online advertising,” Manage. Sci., vol. 67, no. 10, pp. 6030–6054, 2021.
  • [10] B. Van Over, B. Li, E. K. P. Chong, and A. Pezeshki, “An improved greedy curvature bound in finite-horizon string optimization with an application to a sensor coverage problem,” in Proc. 62nd IEEE Conf. Decision Control (CDC), Singapore, Dec. 2023, pp. 1257–1262.
  • [11] M. Streeter and D. Golovin, “An online algorithm for maximizing submodular functions,” in Proc. Adv. in Neural Inf. Process. Syst. 21 (NIPS 2008), vol. 21, Vancouver, BC, Canada, Dec. 2008, pp. 1577–1584.
  • [12] Y. Liu, Z. Zhang, E. K. P. Chong, and A. Pezeshki, “Performance bounds with curvature for batched greedy optimization,” J. Optim. Theory Appl., vol. 177, pp. 535–562, 2018.
  • [13] M. Zhong and C. G. Cassandras, “Distributed coverage control and data collection with mobile sensor networks,” IEEE Trans. Autom. Control, vol. 56, no. 10, pp. 2445–2455, 2011.
  • [14] X. Sun, C. G. Cassandras, and X. Meng, “Exploiting submodularity to quantify near-optimality in multi-agent coverage problems,” Automatica, vol. 100, pp. 349–359, 2019.