This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Energy Consumption of Group Search on a Line 111This is the full version of the paper with the same title which will appear in the proceedings of the 46th International Colloquium on Automata, Languages and Programming 8-12 July 2019, Patras, Greece

Jurek Czyzowicz111 Universite du Québec en Outaouais, Gatineau, Québec, Canada, jurek.czyzowicz@uqo.ca    Konstantinos Georgiou222 Dept. of Mathematics, Ryerson University, Toronto, ON, Canada, konstantinos@ryerson.ca    Ryan Killick444 School of Computer Science, Carleton University, Ottawa ON, Canada, ryankillick,kranakis@scs.carleton.ca    Evangelos Kranakis44footnotemark: 4    Danny Krizanc777 Department of Mathematics & Comp. Sci., Wesleyan University, Middletown, CT, USA, dkrizanc@wesleyan.edu    Manuel Lafond555 Department of Computer Science, Université de Sherbrooke, Québec, Canada Manuel.Lafond@usherbrooke.ca    Lata Narayanan888 Department of Comp. Sci. and Software Eng., Concordia University, Montreal, Québec, Canada, lata,opatrny@encs.concordia.ca    Jaroslav Opatrny88footnotemark: 8    Sunil Shende999 Department of Computer Science, Rutgers University, Camden, USA, shende@camden.rutgers.edu
Abstract

Consider two robots that start at the origin of the infinite line in search of an exit at an unknown location on the line. The robots can collaborate in the search, but can only communicate if they arrive at the same location at exactly the same time, i.e. they use the so-called face-to-face communication model. The group search time is defined as the worst-case time as a function of dd, the distance of the exit from the origin, when both robots can reach the exit. It has long been known that for a single robot traveling at unit speed, the search time is at least 9do(d)9d-o(d); a simple doubling strategy achieves this time bound. It was shown recently in [15] that k2k\geq 2 robots traveling at unit speed also require at least 9d9d group search time.

We investigate energy-time trade-offs in group search by two robots, where the energy loss experienced by a robot traveling a distance xx at constant speed ss is given by s2xs^{2}x and is motivated by first principles in physics and engineering. Specifically, we consider the problem of minimizing the total energy used by the robots, under the constraints that the search time is at most a multiple cc of the distance dd and the speed of the robots is bounded by bb. Motivation for this study is that for the case when robots must complete the search in 9d9d time with maximum speed one (b=1;c=9b=1;~c=9), a single robot requires at least 9d9d energy, while for two robots, all previously proposed algorithms consume at least 28d/328d/3 energy.

When the robots have bounded memory and can use only a constant number of fixed speeds, we generalize an algorithm described in [3, 15] to obtain a family of algorithms parametrized by pairs of b,cb,c values that can solve the problem for the entire spectrum of these pairs for which the problem is solvable. In particular, for each such pair, we determine optimal (and in some cases nearly optimal) algorithms inducing the lowest possible energy consumption.

We also propose a novel search algorithm that simultaneously achieves search time 9d9d and consumes energy 8.42588d8.42588d. Our result shows that two robots can search on the line in optimal time 9d9d while consuming less total energy than a single robot within the same search time. Our algorithm uses robots that have unbounded memory, and a finite number of dynamically computed speeds. It can be generalized for any c,bc,b with cb=9cb=9, and consumes energy 8.42588b2d8.42588b^{2}d.

1 Introduction

The problem of searching for a treasure at an unknown location in a specified continuous domain was initiated over fifty years ago  [6, 7]. Search domains that have been considered include the infinite line [2, 6, 7, 33], a set of rays [10, 11], the unit circle [12, 23, 36], and polygons [26, 32, 34]. Consider a robot (sometimes called a mobile agent) starting at some known location in the domain and looking for an exit that is located at an unknown distance dd away from the start. What algorithm should the robot use to find the exit as soon as possible? The most common cost measure used for the search algorithm is the worst-case search time, as a function of the distance dd of the exit from the starting position. For a fixed-speed robot, the search time is proportional to the length of the trajectory of the robot. Other measures such as turn cost [27] and different costs for revisiting [9] have been sometimes considered.

We consider for the first time the energy consumed by the robots while executing the search algorithm. The energy used by a robot to travel a distance xx at speed ss is computed as s2xs^{2}x and is motivated from the concept of viscous drag in fluid dynamics; see Section 2 for details on the energy model. For a single robot searching on the line, the classic Spiral Search algorithm (also known as the doubling strategy) has search time 9d9d and is known to be optimal when the robot moves with unit speed. Since in the worst case, the robot travels distance 9d9d at unit speed, the energy consumption is 9d9d as well. Clearly, as the speed of the robot increases, the time to find the exit decreases but at the same time, the energy used increases. Likewise, as the speed of the robot decreases, the time to find the exit increases, while the energy consumption decreases. Thus there is a natural trade-off between the time taken by the robot to search for the exit and the energy consumed by the robot. To investigate this trade-off, we consider the problem of minimizing the total energy used by the robots to perform the search when the speed of the robot is bounded by bb, and the time for the search is at most a multiple cc of the distance dd from the starting point to the exit.

Group search by a set of k2k\geq 2 collaborating robots has recently gained a lot of attention. In this case, the search time is the time when all kk robots reach the exit. The problem has also been called evacuation, in view of the application when it is desired that all robots reach and evacuate from the exit. Two models of communication between the robots have been considered. In the wireless communication model, the robots can instantly communicate with each other at any time and over any distance. In the face-to-face communication model (F2F), two robots can communicate only when in the same place at the same time. In many search domains, and for both communication models, group search by k2k\geq 2 agents has been shown to take less time than search by a single agent; see for example [23, 26].

In this paper, we focus on group search on the line, by two robots using the F2F model. Chrobak et al [15] showed that group search in this setting cannot be performed in time less than 9do(d)9d-o(d), regardless of the number of robots, assuming all robots use at most unit speed. They also describe several strategies that achieve search time 9d9d. In the first strategy, the two robots independently perform the Spiral Search algorithm, using unit speed during the entire search. Next, they consider a strategy first described in [3], that we call the Two-Turn strategy, whereby two robots head off independently in opposite directions at speed 1/31/3; when one of them finds the exit, it moves at unit speed to chase and catch the other robot, after which they both return at unit speed to the exit. Finally, they present a new strategy, called the Fast-Slow algorithm in which one robot moves at unit speed, while the other robot moves at speed 1/3, both performing a spiral search. The doubling strategy is very energy-inefficient, it uses energy 18d18d if the two robots always travel together, or 14D14D if the robots start by moving in opposite directions. The other two algorithms both use energy 28d/3>9d28d/3>9d. Interestingly, the two strategies that achieve an energy consumption of 28d/328d/3 with search time 9d9d, both use two different and pre-computed speeds, but are quite different in terms of the robot capacities needed. In the Two-Turn strategy, the robots are extremely simple and use constant memory; they use only three states. In Fast-Slow and Spiral Search, the robots need unbounded memory, and perform computations to determine how far to go before turning and moving in the opposite direction.

Memory capability, time- and speed-bounded search, and energy consumption by a two-robot group search algorithm on the line: these considerations motivate the following questions that we address in our paper:

  1. 1.

    Is there a search strategy for constant-memory robots that has energy consumption <9d<9d?

  2. 2.

    Is there any search strategy that uses time 9d9d and energy <9d<9d?

1.1 Our results

We generalize the Two-Turn strategy for any values of c,bc,b. We analyze the entire spectrum of values of c,bc,b for which the problem admits a solution, and for each of them we provide optimal (and in some cases nearly optimal) speed choices for our robots (Theorem 3.4). In particular, and somewhat surprisingly, our proof makes explicit how for any fixed cc the optimal speed choices do not simply "scale" with bb; rather more delicate speed choices are necessary to comply with the speed and search time bounds. For the special case of cb=9c\cdot b=9, our results match with the specific Two-Turn strategy described in [15]. Our results further show that no Two-Turn strategy can achieve energy consumption less than 9d9d while keeping the search time at 9d9d. In fact, we conjecture that this trade-off is impossible for any group search strategy that uses only constant memory robots.

In the unbounded-memory model, for the special case of c=9c=9 and b=1b=1, we give a novel search algorithm that achieves energy consumption of 8.42588d8.42588d, thus answering the second question above in the affirmative. This result shows that though two robots cannot search faster than one robot on the line [15], somewhat surprisingly, two robots can search using less total energy than one robot, in the same optimal time. Our algorithm uses robots that have unbounded memory, and a finite number of dynamically computed speeds. Note that our algorithm can be generalized for any c,bc,b with cb=9cb=9, and utilizes energy 8.42588b2d8.42588b^{2}d (Theorem 4.7).

1.2 Related Work

Several authors have investigated various aspects of mobile robot (agent) search, resulting in an extensive literature on the subject in theoretical computer science and mathematics (e.g., see [1, 29] for reviews). Search by constant-memory robots has been done mainly for finite-state automata (FSA) operating in discrete environments like infinite grids, their finite-size subsets (labyrinths) and other graphs. The main concern of this research was the feasibility of search, rather than time or energy efficiency. For example, [14] showed that no FSA can explore all labyrinths, while [8] proved that one FSA using two pebbles or two FSAs, communicating according to the F2F model can explore all labyrinths. However, no collection of FSAs may explore all finite graphs communicating in the F2F model [38] or wireless model [17]. On the other hand, all graphs of size nn may be explored using a robot having O(logn)O(\log n) memory [37].

Exploration of infinite grids is known as the ANTS problem [28], where it was shown that four collaborating FSAs in the semi-synchronous execution model and communicating according to the F2F scenario can explore an infinite grid. Recently, [13] showed that four FSAs are really needed to explore the grid (while three FSAs can explore an infinite band of the 2-dimensional grid).

Continuous environment cases have been investigated in several papers when the efficiency of the search is often represented by the time of reaching the target (e.g., see [2, 6, 7, 33]). Even in the case of continuous environment as simple as the infinite line, after the seminal papers [6, 7], various scenarios have been studied where the turn cost has been considered [27], the environment was composed of portions permitting different search speeds [25], some knowledge about the target distance was available [10] or where some other parameters are involved in the computation of the cost function [9] (e.g. when the target is moving).

The group search, sometimes interpreted as the evacuation problem has been studied first for the disc environment under the F2F [12, 18, 23, 26, 35] and wireless [18] communication scenarios and then also for other geometric environments (e.g., see [26]). Other variants of search/evacuation problems with a combinatorial flavour have been recently considered in [16, 19, 20, 30, 31]. Some papers investigated the line search problem in the presence of crash faulty [24] and Byzantine faulty agents [22]. The interested reader may also consult the recent survey [21] on selected search and evacuation topics.

The energy used by a mobile robot is usually considered as being spent solely for travelling. As a consequence, in the case of a single, constant speed robot the search time is proportional to the distance travelled and the energy used by a robot. Therefore the problems of minimization of time, distance or energy are usually equivalent for most robots' tasks. For teams of collaborating robots, the searchers often need to synchronize their walks in order to wait for information communicated by other searchers (e.g, see [12, 18, 35]), hence the time of the task and the distance travelled are different. However, the distance travelled by a robot and its energy used are still commensurable quantities.

To the best of our knowledge, energy consumption as a function of mobile robot speed which is based on natural laws of physics (related to the drag force) has never been studied in the search literature before. Our present work is motivated by [15], which proves that the competitive ratio 99 is tight for group search time with two mobile agents in the F2F model when both agents have unit maximal speeds. More exactly, it follows from [15] that having more unit-speed robots cannot improve the group search time obtained by a single robot. Nevertheless, our paper shows that using more robots can improve the energy spending, while keeping the group-search time still the best possible.

[15] presents interesting examples of group search algorithms for two distinct speed robots communicating according to the F2F scenario. An interested reader may consult  [4], where optimal group search algorithms for a pair of distinct maximal speed robots were proposed for both communication scenarios (F2F and wireless) and for any pair of robots' maximal speeds. It is interesting to note that, according to [4], for any distinct-speed robots with F2F communication, the optimal group search time is obtained only if one of the robots perform the search step not using its full speed.

Paper Organization: In Section 2 we formally define the evacuation problem EEcb\textsc{EE}_{c}^{b}, and proper notions of efficiency. Our algorithms and their analysis for constant-memory robots is presented in Section 3, while in Section 4 we introduce and analyze algorithms for unbounded-memory robots. All omitted proofs can be found in the Appendix. Also, due to space limitations, all figures appear in Appendix A.

2 Preliminaries

Two robots start walking from the origin of an infinite (bidirectional) line in search of a hidden exit at an unknown absolute distance dd from the origin. The exit is considered found only when one of the robots walks over it. An algorithm for group search by two robots specifies trajectories for both robots and terminates when both robots reach the exit. The time by which the second robot reaches the exit is referred to as the search time or the evacuation time.

Robot models: The two robots operate under the F2F communication model in which two robots can communicate only when they are in the same place at the same time. Each robot can change its speed at any time. We distinguish between constant-memory robots that can only travel at a constant number of hard-wired speeds, and unbounded-memory robots that can dynamically compute speeds and distances, and travel at any possible speed.

Energy model: A robot moving at constant speed ss traversing an interval of length xx is defined to use energy s2xs^{2}\cdot x. This model is well motivated from first principles in physics and engineering and corresponds to the energy loss experienced by an object moving through a viscous fluid [5]. In particular, an object moving with constant speed ss will experience a drag force FDF_{D} proportionalThe constant of proportionality has (SI) units kg/mkg/m and depends, among other things, on the shape of the object and the density of the fluid through which it moves. to s2s^{2}. In order to maintain the speed ss over a distance xx the object must do work equal to the product of FDF_{D} and xx resulting in a continuous energy loss proportional to the product of the object's squared speed and travel distance. For simplicity we have taken the proportionality constant to be one.

The total energy that a robot uses traveling at speeds s1,s2,,sts_{1},s_{2},\ldots,s_{t}, traversing intervals x1,x2,,xtx_{1},x_{2},\ldots,x_{t}, respectively, is defined as i=1tsi2xi\sum_{i=1}^{t}s_{i}^{2}\cdot x_{i}. For group search with two robots, the energy consumption is defined as the sum total of the two robots' energies used until the search algorithm terminates.

For each d>0d>0 there are two possible locations for the exit to be at distance dd from the origin: we will refer to either of these as input instances dd for the group search problem. Our goal is to solve the following optimized search problem parametrized by two values, 𝐛\mathbf{b} and 𝐜\mathbf{c}:

\blacktriangleright Problem EEcb\textsc{EE}_{c}^{b}: Design a group search algorithm for two robots in the F2F model that minimizes the energy consumption for dd-instances under the constraints that the search time is no more than 𝐜𝐝\mathbf{c\cdot d} and the robots use speeds that are at most 𝐛\mathbf{b}. When there are no speed limits on the robots (i.e. b=b=\infty), we abbreviate EEc\textsc{EE}_{c}^{\infty} by EEc\textsc{EE}_{c}. Note that b,cb,c are inputs to the algorithm, but dd and the exact location of the exit are not known.

As it is standard in the literature on related problems, we assume that the exist is at least a known constant distance away from the origin. In this work, we pick the constant equal to 2, although our arguments can be adjusted to any other constant. It is not difficult to show that EEcb\textsc{EE}_{c}^{b} is well defined for each b,c>0b,c>0 with bc1bc\geq 1, and the optimal offline solution, for instance dd, is for both robots to move at speed 1c\frac{1}{c} to the exit. This offline algorithm has energy consumption 2dc2\frac{2d}{c^{2}} (see Observation B.1 in Appendix B). Consider an online algorithm for EEcb\textsc{EE}_{c}^{b}, which on any instance dd has energy consumption at most e(c,b,d)e(c,b,d). The competitive ratio of the algorithm is defined as supd>0c22de(c,b,d).\sup_{d>0}\tfrac{c^{2}}{2d}~e(c,b,d).

Due to [15], and when b=1b=1, no online algorithm (for two robots) can have evacuation time less than 9dϵ9d-\epsilon (for any ϵ>0\epsilon>0 and for large enough dd). By scaling, using arbitrary speed limit bb, we obtain the following fact.

Observation 2.1.

No online F2F algorithm can solve EEcb\textsc{EE}_{c}^{b} if cb<9cb<9.

3 Solving EEcb\textsc{EE}_{c}^{b} with Constant-Memory Robots

In this section we propose a family of algorithms for solving EEcb\textsc{EE}_{c}^{b} (including b=b=\infty). The family uses an algorithm that is parametrized by three discrete speeds: ss, rr and kk. The robots use these speeds depending on finite state control as follows:

\blacktriangleright Algorithm 𝒩s,r,k\mathcal{N}_{s,r,k}: Robots start moving in opposite directions with speed ss until the exit is found by one of them. The finder changes direction and moves at speed r>sr>s until it catches the other robot. Together the two robots return to the exit using speed kk.

Lemma 3.1 (Proof on page C.1).

Let b,cb,c be such that there exist s,r,ks,r,k for which 𝒩s,r,k\mathcal{N}_{s,r,k} is feasible. Then, for instance dd of EEcb\textsc{EE}_{c}^{b}, the induced evacuation time of 𝒩s,r,k\mathcal{N}_{s,r,k} is dT(s,r,k)d\cdot T\left({s,r,k}\right) and the induced energy consumption is 2dE(s,r,k)2d\cdot E\left({s,r,k}\right), where

T(s,r,k):=2(k+r)k(rs)+1s,E(s,r,k):=rrs(s2+r2+2k2)T\left({s,r,k}\right):=\frac{2(k+r)}{k(r-s)}+\frac{1}{s},~~E\left({s,r,k}\right):=\frac{r}{r-s}\left(s^{2}+r^{2}+2k^{2}\right)

We propose a systematic way in order to find optimal values for s,r,ks,r,k of algorithm 𝒩s,r,k\mathcal{N}_{s,r,k} for optimization problem EEcb\textsc{EE}_{c}^{b} (including b=b=\infty), whenever such values exist.

Theorem 3.2 (Proof on page C.2).

Algorithm 𝒩s,r,k\mathcal{N}_{s,r,k} gives rise to a feasible solution to problem EEcb\textsc{EE}_{c}^{b} if and only if bc9bc\geq 9. For every such b,c>0b,c>0, the optimal choices of 𝒩s,r,k\mathcal{N}_{s,r,k} can be obtained by solving Non Linear Program:

mins,r,kE(s,r,k)\displaystyle~~\min_{s,r,k\in\mathbb{R}}E\left({s,r,k}\right) (NLPcb\textsc{NLP}_{c}^{b})
s.t.\displaystyle s.t.~~ T(s,r,k)c\displaystyle T\left({s,r,k}\right)\leq c
rs\displaystyle r\geq s
0s,r,kb\displaystyle 0\leq s,r,k\leq b

where functions E(,,),T(,,)E\left({\cdot,\cdot,\cdot}\right),T\left({\cdot,\cdot,\cdot}\right) are as in Lemma 3.1. Moreover, if s0,r0,k0s_{0},r_{0},k_{0} are the optimizers to NLPcb\textsc{NLP}_{c}^{b}, then the competitive ratio of 𝒩s0,r0,k0\mathcal{N}_{s_{0},r_{0},k_{0}} equals c2E(s0,r0,k0).c^{2}\cdot E\left({s_{0},r_{0},k_{0}}\right).

The following subsections are devoted to solving NLPcb\textsc{NLP}_{c}^{b}, effectively proving Theorem 3.4. First in Section 3.1 we solve the case b=b=\infty and we use our findings to solve the case of bounded speeds bb in the follow-up Section 3.2.

3.1 Optimal Choices of 𝒩s,r,k\mathcal{N}_{s,r,k} for the Unbounded-Speed Problem

In this section we propose solutions to the unbounded-speed problem EEc\textsc{EE}_{c}. Since EEc\textsc{EE}_{c} is the same as EEc\textsc{EE}_{c}^{\infty}, by Observation B.1, the problem is well-defined for every fixed c>0c>0. Moreover, by the proof of Theorem 3.2 (see proof of Lemma C.1 in the Appendix) algorithm 𝒩s,r,k\mathcal{N}_{s,r,k} induces a feasible solution for every c>0c>0 as well, and the optimal speeds can be found by solving NLPc\textsc{NLP}_{c}^{\infty}. Indeed, in the remaining of the section we show how to choose optimal values for s,r,ks,r,k for solving EEc\textsc{EE}_{c} with 𝒩s,r,k\mathcal{N}_{s,r,k}. Let

σ2.65976,ρ11.3414,κ6.63709,\sigma\approx 2.65976,\rho\approx 11.3414,\kappa\approx 6.63709, (1)

whose exact values are the roots of an algebraic system and will be formally defined later. The main theorem of this section reads as follows.

Theorem 3.3 (Proof on page C.3).

Let σ,ρ,κ\sigma,\rho,\kappa as in (1). For every c>0c>0, the optimal speeds of 𝒩s,r,k\mathcal{N}_{s,r,k} for problem EEc\textsc{EE}_{c} are s=σc,r=ρc,k=κc.s=\tfrac{\sigma}{c},~~r=\tfrac{\rho}{c},~~k=\tfrac{\kappa}{c}. Moreover, the competitive ratio of the corresponding solution is independent of cc and equals ρ(2κ2+ρ2+σ2)ρσ292.369\frac{\rho\left(2\kappa^{2}+\rho^{2}+\sigma^{2}\right)}{\rho-\sigma}\approx 292.369.

A high level outline of the proof of Theorem 3.3 is as follows. First we show that any optimal choices of the speeds of 𝒩s,r,k\mathcal{N}_{s,r,k} must satisfy the time constraint of NLPc\textsc{NLP}_{c}^{\infty} tightly. Then, we show that finding optimal speeds s,r,ks,r,k of 𝒩s,r,k\mathcal{N}_{s,r,k} for the general problem EEc\textsc{EE}_{c} reduces to problem EE1\textsc{EE}_{1}. Finally, we obtain the optimal solution to NLP1\textsc{NLP}_{1}^{\infty} by standard tools of nonlinear programming (KKT conditions).

3.2 (Sub)Optimal Choices of 𝒩s,r,k\mathcal{N}_{s,r,k} for the Bounded-Speed Problem

In this section, we show how to choose optimal values for s,r,ks,r,k for solving EEcb\textsc{EE}_{c}^{b} with 𝒩s,r,k\mathcal{N}_{s,r,k}, for the entire spectrum of c,bc,b values for which the problem is solvable by online algorithms.

The main result of this section is the following:

Theorem 3.4.

Let γ19.06609\gamma_{1}\approx 9.06609, γ2=ρ11.3414\gamma_{2}=\rho\approx 11.3414, and σ,ρ,κ\sigma,\rho,\kappa as in (1). For every c,b>0c,b>0 with cb9cb\geq 9, the following choices of speeds s,r,ks,r,k are feasible for 𝒩s,r,k\mathcal{N}_{s,r,k}

9cbγ19\leq cb\leq\gamma_{1} γ1<cb<γ2\gamma_{1}<cb<\gamma_{2} cbγ2cb\geq\gamma_{2}
ss (bc)210bc+9+bc32c\frac{-\sqrt{(bc)^{2}-10bc+9}+bc-3}{2c} 0.532412b0.0262661b2c0.532412b-0.0262661b^{2}c σ/c\sigma/c
rr bb bb ρ/c\rho/c
kk bb 2bsbcsbcs2s\frac{2bs}{bcs-b-cs^{2}-s} κ/c\kappa/c

The induced competitive ratio is given by:

f(x):={12x(x(x(x9)(x1))+(x9)(x1)+3),9xγ1x2((0.5324120.0262661x)2+11595.8(20.26991.x)2(x(x(x2.46798)398.916)+2221.18)2+1)0.0262661x+0.467588,γ1<x<γ2292.369xγ2f(x):=\left\{\begin{array}[]{ll}\frac{1}{2}x\left(x\left(x-\sqrt{(x-9)(x-1)}\right)+\sqrt{(x-9)(x-1)}+3\right),&9\leq x\leq\gamma_{1}\\ \frac{x^{2}\left((0.532412\,-0.0262661x)^{2}+\frac{11595.8(20.2699\,-1.x)^{2}}{(x(x(x-2.46798)-398.916)+2221.18)^{2}}+1\right)}{0.0262661x+0.467588},&\gamma_{1}<x<\gamma_{2}\\ 292.369&x\geq\gamma_{2}\end{array}\right.

and the induced energy, for instances dd, is f(cb)2dc2f(cb)\frac{2d}{c^{2}}. Moreover, the competitive ratio depends only on the product cbcb.

In particular, the speeds' choices are optimal when cbγ1cb\leq\gamma_{1} and when cbγ2cb\geq\gamma_{2}. When γ1<cb<γ2\gamma_{1}<cb<\gamma_{2}, the derived competitive ratio is no more than 0.03 additively off from that induced by optimal choices of s,r,ks,r,k.

Corollary 3.5.

For c=9,b=1c=9,b=1, the bounded-memory robot algorithm 𝒩s,r,k\mathcal{N}_{s,r,k} has energy consumption 28d/328d/3 and competitive ratio 378.

Theorem 3.4 is proven by solving NLPcb\textsc{NLP}_{c}^{b} of Theorem 3.2. In particular, the induced competitive ratio of 𝒩s,r,k\mathcal{N}_{s,r,k} for the choices of Theorem 3.4 is summarized in Figure 1 (Appendix A). Speed values s,r,ks,r,k, are chosen optimally when cbcb is either at most γ1\gamma_{1} or at least γ2\gamma_{2} (i.e. optimizers to NLPcb\textsc{NLP}_{c}^{b} admit analytic description). The optimal speed parameters when γ1<cb<γ2\gamma_{1}<cb<\gamma_{2} cannot be determined analytically (they are roots of high degree polynomials). The values that appear in Theorem 3.4 are heuristically chosen, but interestingly induce nearly optimal competitive ratio, see Figure 2 (Appendix A).

The proof of Theorem 3.4 is given by Lemma 3.6 (the case cbγ1cb\leq\gamma_{1}), Lemma 3.7 (the case cbγ2cb\geq\gamma_{2}), and Lemma 3.8 (the case γ1<cb<γ2\gamma_{1}<cb<\gamma_{2}). Next we state these Lemmata, and we sketch their proofs.

Lemma 3.6 (Proof on page C.4).

For every c(9/b,γ1/b]c\in(9/b,\gamma_{1}/b], where γ19.06609\gamma_{1}\approx 9.06609, the optimizers to NLPcb\textsc{NLP}_{c}^{b} are k=r=bk=r=b, and sb=(bc)210bc+9+bc32cs_{b}=\frac{-\sqrt{(bc)^{2}-10bc+9}+bc-3}{2c}. The induced competitive ratio is f(cb)f(cb), (see definition of f(x)f(x) for xγ1x\leq\gamma_{1} in statement of Theorem 3.4), and the energy consumption, for instances dd, is f(cb)2dc2f(cb)\frac{2d}{c^{2}}.

For proving Lemma 3.6, first we recall the known optimizer for the special case cb=9cb=9 (see Corollary C.6 within the Proof of Lemma 3.6 on page C.4), and we identify the tight constraints. Requiring that the exact same inequality constraints to NLPcb\textsc{NLP}_{c}^{b} remain tight, we ask how large can the product cbcb be so as to have KKT condition hold true. From the corresponding algebraic system, we obtain the answer cbγ19.06609cb\leq\gamma_{1}\approx 9.06609.

Similarly, from Theorem 3.3 we know the optimizers to NLPcb\textsc{NLP}_{c}^{b} for large enough values of cbcb, and the corresponding tight constraints to the NLP. Again, using KKT conditions, we show that the same constraints remain tight for the optimizers as long as cbγ211.3414cb\geq\gamma_{2}\approx 11.3414. This way we obtain the following Lemma.

Lemma 3.7 (Proof on page C.5).

For every c>ρ/b11.3414/bc>\rho/b\approx 11.3414/b, the optimal speeds of 𝒩s,r,k\mathcal{N}_{s,r,k} for EEcb\textsc{EE}_{c}^{b} are s=σ/c,r=ρ/c,k=κ/c,s=\sigma/c,~r=\rho/c,~k=\kappa/c, i.e. they are the same as for EEc\textsc{EE}_{c}^{\infty}. If the target is placed at distance dd from the origin, then the induced energy equals 584.738dc2584.738\frac{d}{c^{2}}. Moreover, the induced competitive ratio is 292.369292.369, and is independent of b,cb,c.

The case γ1<cb<γ2\gamma_{1}<cb<\gamma_{2} can be solved optimally only numerically, since the best speed values are obtained by roots to a high degree polynomial. Nevertheless, the following lemma proposes a heuristic choice of speeds (that of Theorem 3.4) which is surprisingly close to the optimal (as suggested by Theorem 3.4, see also Figure 2).

Lemma 3.8 (Proof on page C.6).

The choices of s,r,ks,r,k of Theorem 3.4 when γ1<cb<γ2\gamma_{1}<cb<\gamma_{2} are feasible. Moreover, the induced competitive ratio is at most 0.03 additively off from the competitive ratio induced by the optimal choices of speeds (evaluated numerically).

The trick in order to find ``good enough'' optimizers to NLPcb\textsc{NLP}_{c}^{b} is to guess the subset of inequality constraints that remain tight when γ1<cb<γ2\gamma_{1}<cb<\gamma_{2}. First, we observe that constraint rbr\leq b is tight for the provable optimizers for all c,bc,b when cb[9,γ1][γ2,)cb\in[9,\gamma_{1}]\cup[\gamma_{2},\infty). As the only other constraint that switches from being tight to non-tight in the same interval is kbk\leq b, we are motivated to maintain tightness for constraints rbr\leq b and the time constraint. Still the algebraic system associated with the corresponding KKT conditions cannot be solved analytically. To bypass this difficulty, and assuming we know (optimal) speed ss, we use the tight time constraint to find speed kk as a function of c,b,sc,b,s. From numerical calculations, we see that optimal speed ss is nearly optimal in cc, and so we heuristically set s=αc+βs=\alpha c+\beta. We choose α,β\alpha,\beta so as to have ss satisfy optimality conditions for the boundary values cb=γ1,γ2cb=\gamma_{1},\gamma_{2}. After we identify all parameters to our solution, we compare the value of our solution to the optimal one (obtained numerically), and we verify (using numerical calculations) that our heuristic solution is only by at most 0.03 additively off. The advantage of our analysis is that we obtain closed formulas for the speed parameters for all values of cb9cb\geq 9.

4 Solving EEcb\textsc{EE}_{c}^{b} with Unbounded-Memory Robots

In this section we prove Theorem 4.7, that is we solve EEcb\textsc{EE}_{c}^{b} by assuming that the two robots have unbounded memory, and in particular that they can perform time and state dependent calculations and tasks. Note that, by scaling, our results hold for all b,cb,c for which cb=9cb=9. For simplicity our exposition is for the natural case c=9c=9 and b=1b=1. Also, as before, dd will denote the unknown distance of the exit from the origin, still the exit is assumed, for the purposes of performance analysis, to be at least 2 away from the origin.

Throughout the execution of our evacuation algorithm, robots can be in 3 different states (similar to the case of constant-memory robots). First, both robots start with the Exploration State and they remain in this until the exit is located. While in the exploration state, robots execute an elaborate exploration that requires synchronous movements in which robots, at a high level, stay in good proximity, still they expand the searched space relatively fast. Then, the exit finder enters the Chasing State in which the robot, depending on its distance from the origin, calculates a speed, at which to move in order to catch and notify the other robot. Lastly, when the two robots meet, they both enter the Exit State in which both robots move toward the exit with the smallest possible speed while meeting the time constraint.

Our algorithm takes as input the values of c=9,b=1c=9,b=1, and use a speed value sbs\leq b, that will be chosen later. When the exit finder switches its state from Exploration to Chasing, it remembers the distance dd of the exit to the origin, as well as the value kk of a counter that was used while in the Exploration State. When the exit finder catches the other robot, they both switch to the Exit State, and they remember their distance pp from the origin, as well as the value of time tt that their rendezvous was realized. The speed of their Exit State will be determined as a function of p,d,tp,d,t (and hence of s,c,bs,c,b as well).

4.1 A Critical Component: ll-Phase Explorations

We adopt the language of [15] in order to discuss a structural property that any feasible evacuation algorithm for EE91\textsc{EE}_{9}^{1} satisfies. As a result, the purpose of this section is to provide high level intuition for our evacuation algorithm that is presented in subsequent sections.

We refer to the two robots (starting exploration from the origin) as LL and RR, intended to explore to the left and to the right of the origin, respectively. The robot trajectories can be drawn on the Cartesian plane where point-location (x,t)(x,-t) will correspond to point xx on the line being visited by some robot at time tt. The following Theorem is due to [15] and was originally phrased for the time-evacuation unit-speed robots' problem. We adopt the language of our problem.

Theorem 4.1.

For any feasible solution to EE91\textsc{EE}_{9}^{1}, the point-location of any robot lies within the cone spanned by vectors (13),(13)\binom{-1}{-3},\binom{1}{-3}.

Next we present some preliminaries toward describing our kk-phase exploration algorithms. A phase is a pair (s,r)(s,r) where s[0,1]s\in[0,1] is a speed and rr\in\mathbb{R} is a distance ratio, possibly negative. An ll-phase algorithm is determined by a position p0p_{0} on the line and a sequence S=(s1,r1),,(sk,rl)S=(s_{1},r_{1}),\ldots,(s_{k},r_{l}) of ll phases (movement instructions). Whenever rix<0r_{i}x<0, movement will be to the left, whereas rix>0r_{i}x>0 will correspond to movement to the right.

ll-phase Exploration: given p0p_{0} and S=(s1,r1),,(sl,rl)S=(s_{1},r_{1}),\ldots,(s_{l},r_{l})
Go to p0p_{0} at speed 1/31/3
repeat
 xx\leftarrow current position
 for i=1,,li=1,\ldots,l do
      Travel at speed sis_{i} for a distance of rixr_{i}\cdot x
   end for
 
end

We will make sure that each time the loop is executed, position xx and corresponding time induce point-locations of the robots that lie in the boundary of the cone of Theorem 4.1. If a loop starts at location xx, then it takes additional time i[l]|ri||x|si\sum_{i\in[l]}\frac{|r_{i}||x|}{s_{i}} to complete one iteration. We will be referring to quantity 1+i[l]|ri|3si1+\sum_{i\in[l]}\frac{|r_{i}|}{3s_{i}} as the expansion factor of Exploration SS.

4.2 Algorithm 𝒜(s)\mathcal{A}\left(s\right): The Exploration, Chasing and Exit States

In this section we give a formal description of our evacuation algorithm. The most elaborate part of it is when robots are in Exploration States, in which they will perform 33-phase exploration. It can be shown that 33-phase exploration based evacuation algorithms that do not violate the constraints of problem EE91\textsc{EE}_{9}^{1} have expansion factor at most 4. Moreover, among those, the ones who minimize the induced energy consumption energy consumption make robots move at speed 1 in the first and third phaseThe proof of these facts is lengthy and technical, and is not required for the correctness of our algorithm, rather it only justifies some parameter choices. Robot's speed in the second phase will be denoted by ss.

We now present a specific 33-phase exploration algorithm, that we denote by 𝒜(s)\mathcal{A}\left(s\right), complying with the above conditions, with phases (1,1),(4s/(1s),s)(-1,1),(4s/(1-s),s) and (44s/(1s),1)(4-4s/(1-s),1), where ss is an exploration speed to be determined later. Robot LL will execute the 3-phase exploration with starting position -1, while robot RR with starting position 22. When subroutine travel(v,p)travel(v,p) is invoked, the robot sets its speed to vv and, from its current position, goes toward position pp on the line until it reaches it. We depict the trajectories of the robots while in the Exploration State in Figure 3.

Exploration State of LL travel(1/3,1)travel(1/3,-1) k0k\leftarrow 0 repeat  travel(1,0)travel(1,0)  travel(s,4k+1s1s)travel(s,-4^{k+1}\cdot\frac{s}{1-s})  travel(1,4k+1)travel(1,-4^{k+1})  kk+1k\leftarrow k+1 end Exploration State of RR travel(1/3,2)travel(1/3,2) k0k\leftarrow 0 repeat  travel(1,0)travel(1,0)  travel(s,24k+1s1s)travel(s,2\cdot 4^{k+1}\cdot\frac{s}{1-s})  travel(1,24k+1)travel(1,2\cdot 4^{k+1})  kk+1k\leftarrow k+1 end

A complete execution of one repeat loop within the Exploration State will be referred to as a round. Variable kk counts the number of completed rounds. Each robot stays in the Exploration State till the exit it found. When switching to the Chasing state (which happens only for the exit finder), robot remembers its current value of counter kk, as well as the distance dd of the exit to the origin. Based on these values (as well as ss) it calculates the most efficient trajectory in order to catch the other robot (predicting, when applicable, that the rendezvous can be realized while the other robot is approaching the exit finder). When the rendezvous is realized, robots store their current distance pp to the origin, as well as the time tt that has already passed. Then, robots need to travel distance p+dp+d to reach the exit. Knowing they have time 9dt9d-t remaining, they go to the exit together as slow as possible to reach the exit in time exactly 9d9d. Figure 4 provides an illustration of the behavior of the robots after finding the exit.

Chasing State K4kK\leftarrow 4^{k} if I am RR then  K24kK\leftarrow 2\cdot 4^{k} end if smin{d4Kd/s,1}s^{\prime}\leftarrow\min\left\{\dfrac{d}{4K-d/s},1\right\} Travel toward the other robot at speed ss^{\prime} until meeting it at distance pp from the origin, and at time tt. Exit State s¯p+d9dt\bar{s}\leftarrow\frac{p+d}{9d-t} Go toward the exit with speed s¯\bar{s}.

4.3 Performance Analysis & an Optimal Choice for Parameter ss

In this section we are ready to provide the details for proving Theorem 4.7. Evacuation algorithm 𝒜(s)\mathcal{A}\left(s\right) is not feasible to EEcb\textsc{EE}_{c}^{b} for all values of speed parameter ss (of the Exploration States). We will show later that trajectories induce evacuation time at most 9d9d only if s[1/3,1/2]s\in[1/3,1/2]. In what follows, and even though we have not fixed the value of ss yet, we will assume that ss has some value between 1/3 and 1/2. The purpose of this section is to fix a value for parameter ss, show that 𝒜(s)\mathcal{A}\left(s\right) is feasible to EE91\textsc{EE}_{9}^{1}, and subsequently compute the induced energy consumption and competitive ratio. As a reminder, each iteration of the repeat loop of the Exploration States is called a round, and kk is a counter for these rounds.

Proposition 4.2 (Proof on page D.1).

For every k0k\geq 0, and at the start of its kk-th round,
\bullet robot LL is at position 4k-4^{k} at time 34k3\cdot 4^{k}, and
\bullet robot RR, is at position 24k2\cdot 4^{k} at time 64k6\cdot 4^{k}.

Let X{L,R}X\in\{L,R\} be one of the robots. We define K(X,k)=4kK(X,k)=4^{k} if X=LX=L, and K(X,k)=24kK(X,k)=2\cdot 4^{k} if X=RX=R, i.e. the position of XX at the start of round kk. We will often analyze 3 cases for the distance dd of the exit with respect to K:=K(X,k)K:=K(X,k) (as it also appears in the description of the Chasing State), associated with the following closed intervals

D1(K):=[K,4Ks/(s+1)],D2(K):=[4Ks/(s+1),4Ks/(1s)],D3(K):=[4Ks/(1s),4K].D_{1}(K):=[K,4Ks/(s+1)],~D_{2}(K):=[4Ks/(s+1),4Ks/(1-s)],~D_{3}(K):=[4Ks/(1-s),4K].

We may simply write D1,D2D_{1},D_{2} and D3D_{3} if KK is clear from the context. Note that during the second phase of round KK, robot LL explores D1D_{1} and D2D_{2}, whereas D3D_{3} is explored during the third phase. The same statement holds for RR. The following lemma will be useful in analyzing the worst case evacuation time and energy consumption of our algorithm.

Lemma 4.3 (Proof on page D.2).

Suppose that robot X{L,R}X\in\{L,R\} finds the exit at distance dd when its round counter has value kk. Let pp and tt be, respectively, the position and time at which XX first meets with the other robot after having found the exit, and set K:=K(X,k)K:=K(X,k). Then the following hold:

  1. 1.

    If dD1d\in D_{1}, then p=0p=0 and t=8Kt=8K.

  2. 2.

    If dD2d\in D_{2}, then |p|=d+ds4Ks1s|p|=\frac{d+ds-4Ks}{1-s} and t=8K+d+d/s4K1st=8K+\frac{d+d/s-4K}{1-s}.

  3. 3.

    If dD3d\in D_{3}, then |p|=2ds/(1s)|p|=2ds/(1-s) and t=8K+2d+2ds/(1s)t=8K+2d+2ds/(1-s).

Using the lemma above, we can now prove that 𝒜(s)\mathcal{A}\left(s\right) meets the speed bound and the evacuation time bound.

Lemma 4.4 (Proof on page D.3).

For any s[1/3,1/2]s\in[1/3,1/2], evacuation algorithm 𝒜(s)\mathcal{A}\left(s\right) is feasible to EE91\textsc{EE}_{9}^{1}.

Lemma 4.3 allows us to derive the speed sb1,sb2s_{b1},s_{b2} and sb3s_{b3} at which both robots go toward the exit after meeting for the cases dD1,dD2d\in D_{1},d\in D_{2} and dD3d\in D_{3}, respectively. We also know the speed sc1s_{c1} at which the exit-finder catches up to the other robot when dD1d\in D_{1}. We define

sb1:=d9d8K,sc1=d4Kd/s,sb2:=2d4Ksd(89s1/s)+4K(2s1),sb3:=d(1+s)d(79s)+8K(s1)s_{b1}:=\tfrac{d}{9d-8K},~~s_{c1}=\tfrac{d}{4K-d/s},~~s_{b2}:=\tfrac{2d-4Ks}{d(8-9s-1/s)+4K(2s-1)},~~s_{b3}:=\tfrac{d(1+s)}{d(7-9s)+8K(s-1)}

The speed sb2s_{b2} is a simple rearrangement of the speed d+qs9d(8K+q), where q=d+d/s4K1s\frac{d+qs}{9d-(8K+q)},\mbox{ where }q=\frac{d+d/s-4K}{1-s}, and sb3s_{b3} is obtained by rearranging d+2ds/(1s)9d(8K+2d+2ds/(1s))\frac{d+2ds/(1-s)}{9d-(8K+2d+2ds/(1-s))}.

Next compute the energy consumption. For given K,dK,d and ss, denote by EL(K,d,s)E_{L}(K,d,s) the energy spent by robot LL from time 33 to time 9d9d when it exits. Similarly, ER(K,d,s)E_{R}(K,d,s) is the energy spent by RR from time 66 to time 9d9d. Then, then energy consumption is E(K,d,s):=13+EL(K,d,s)+ER(K,d,s)E(K,d,s):=\tfrac{1}{3}+E_{L}(K,d,s)+E_{R}(K,d,s). For any KK and ss, we also define F(K,s):=(K1)(54s(s+1))F(K,s):=(K-1)(5-4s(s+1)).

Lemma 4.5 (Proof on page D.4).

Suppose that robot X{L,R}X\in\{L,R\} finds the exit at distance dd when its round counter has value kk, and let K:=K(X,k)K:=K(X,k). Then

E(K,d,s)=13+{F(K,s)+3K+d(s2+sc12+2sb12)if dD1F(K,s)+3K+(2d4Ks1s)(1+s2+2sb22)if dD2F(K,s)+3K4Ks(s+1)+2d1s(s3+sb32(s+1)+1)if dD3.E(K,d,s)=\frac{1}{3}+\begin{cases}F(K,s)+3K+d(s^{2}+s_{c1}^{2}+2s_{b1}^{2})&\mbox{if }d\in D_{1}\\ F(K,s)+3K+\left(\dfrac{2d-4Ks}{1-s}\right)(1+s^{2}+2s_{b2}^{2})&\mbox{if }d\in D_{2}\\ F(K,s)+3K-4Ks(s+1)+\dfrac{2d}{1-s}(s^{3}+s_{b3}^{2}(s+1)+1)&\mbox{if }d\in D_{3}.\end{cases}

Denote by Ei(k,d,s)E_{i}(k,d,s) the value of E(K,d,s)E(K,d,s) when dDid\in D_{i}, i=1,2,3i=1,2,3. Our intension now is to fix speed value ss that solves the following Nonlinear Program

mins[1/3,1/2]{max{supdD1,k1,XE1(K,d,s)d,supdD2,k1,XE2(K,d,s)d,supdD3,k1,XE3(K,d,s)d}}.\min_{s\in[1/3,1/2]}\left\{\max\left\{\sup_{d\in D_{1},k\geq 1,X}\frac{E_{1}(K,d,s)}{d},\sup_{d\in D_{2},k\geq 1,X}\frac{E_{2}(K,d,s)}{d},\sup_{d\in D_{3},k\geq 1,X}\frac{E_{3}(K,d,s)}{d}\right\}\right\}.

For every s[1/3,1/2]s\in[1/3,1/2] we show in Lemma 4.6 that E1(K,d,s)d\frac{E_{1}(K,d,s)}{d} is decreasing in dD1d\in D_{1}, that E2(K,d,s)d\frac{E_{2}(K,d,s)}{d} is increasing in dD2d\in D_{2}, and that E3(K,d,s)d\frac{E_{3}(K,d,s)}{d} is decreasing in dD3d\in D_{3}. Then, the best parameter ss can be chosen so as to make all worst case valued Ei(K,d,s)d\frac{E_{i}(K,d,s)}{d} equal (if possible) when i=1,2,3i=1,2,3. The optimal ss can be found by numerically finding the roots of a high degree polynomial, and accordingly, we heuristically set s=0.39403s=0.39403, inducing the best possible energy consumption for algorithm 𝒜(s)\mathcal{A}\left(s\right). All relevant formal arguments are within the proof of the next lemma.

Lemma 4.6 (Proof on Page D.5).

On instance dd of EE91\textsc{EE}_{9}^{1}, algorithm 𝒜(s)\mathcal{A}\left(s\right) induces energy consumption at most 8.42588d8.42588d, when s=0.39403s=0.39403.

By Lemma 4.6, we conclude that for the specific value of ss, algorithm 𝒜(s)\mathcal{A}\left(s\right) has competitive ratio 9228.42588341.24814,\frac{9^{2}}{2}8.42588\approx 341.24814, concluding the proof of Theorem 4.7.

Theorem 4.7.

For every c,b>0c,b>0 with cb=9cb=9, there is an evacuation algorithm for unbounded-memory autonomous robots solving EEcb\textsc{EE}_{c}^{b} inducing energy consumption 8.42588b2d8.42588b^{2}d for instances dd, and competitive ratio 341.24814.

Acknowledgements

Research supported by NSERC discovery grants, NSERC graduate scholarship, and NSF.

References

  • [1] S. Alpern and S. Gal. The theory of search games and rendezvous. Springer, 2003.
  • [2] R. Baeza Yates, J. Culberson, and G. Rawlins. Searching in the plane. Information and Computation, 106(2):234–252, 1993.
  • [3] R. Baeza-Yates and R. Schott. Parallel searching in the plane. Computational Geometry, 5(3):143–154, 1995.
  • [4] E. Bampas, J. Czyzowicz, L. Gasieniec, D. Ilcinkas, R. Klasing, T. Kociumaka, and D. Pajak. Linear search by a pair of distinct-speed robots. Algorithmica, 81(1):317–342, 2019.
  • [5] G. K. Batchelor. An Introduction to Fluid Dynamics. Cambridge Mathematical Library. Cambridge University Press, 2000.
  • [6] A. Beck. On the linear search problem. Israel J. of Mathematics, 2(4):221–228, 1964.
  • [7] R. Bellman. An optimal search. SIAM Review, 5(3):274–274, 1963.
  • [8] M. Blum and D. Kozen. On the power of the compass (or, why mazes are easier to search than graphs). In FOCS, pages 132–142, 1978.
  • [9] P. Bose and J.-L. De Carufel. A general framework for searching on a line. Theoretical Computer Science, pages 703:1–17, 2017.
  • [10] P. Bose, J.-L. De Carufel, and S. Durocher. Searching on a line: A complete characterization of the optimal solution. Theoretical Computer Science, pages 569:24–42, 2015.
  • [11] S. Brandt, K.-T. Foerster, B. Richner, and R. Wattenhofer. Wireless evacuation on m rays with k searchers. In SIROCCO, pages 140–157, 2017.
  • [12] S. Brandt, F. Laufenberg, Y. Lv, D. Stolz, and R. Wattenhofer. Collaboration without communication: Evacuating two robots from a disk. In CIAC, pages 104–115, 2017.
  • [13] S. Brandt, J. Uitto, and R. Wattenhofer. A tight lower bound for semi-synchronous collaborative grid exploration. In DISC, pages 13:1–13:17, 2018.
  • [14] L. Budach. Automata and labyrinths. Math. Nachrichten, 86:195–282, 1978.
  • [15] M. Chrobak, L. Gasieniec, Gorry T., and R. Martin. Group search on the line. In SOFSEM, pages 164–176. Springer, 2015.
  • [16] H. Chuangpishit, K. Georgiou, and P. Sharma. Average case - worst case tradeoffs for evacuating 2 robots from the disk in the face-to-face model. In ALGOSENSORS'18. Springer, 2018.
  • [17] S. A. Cook and C. Rackoff. Space lower bounds for maze threadability on restricted machines. SIAM Journal on Computing, 9(3):636–652, 1980.
  • [18] J. Czyzowicz, L. Gasieniec, T. Gorry, E. Kranakis, R. Martin, and D. Pajak. Evacuating robots via unknown exit in a disk. In DISC, pages 122–136. Springer, 2014.
  • [19] J. Czyzowicz, K. Georgiou, R. Killick, E. Kranakis, D. Krizanc, L. Narayanan, J. Opatrny, and S. Shende. God save the queen. In (FUN), pages 16:1–16:20, 2018.
  • [20] J. Czyzowicz, K. Georgiou, R. Killick, E. Kranakis, D. Krizanc, L. Narayanan, J. Opatrny, and S. Shende. Priority evacuation from a disk using mobile robots. In SIROCCO, pages 209–225, 2018.
  • [21] J. Czyzowicz, K. Georgiou, and E. Kranakis. Group search and Evacuation. In Distributed Computing by Mobile Entities, Current Research in Moving and Computing, LNCS, volume 11340, pages 335–370, 2019.
  • [22] J. Czyzowicz, K. Georgiou, E. Kranakis, D. Krizanc, L. Narayanan, J. Opatrny, and S. Shende. Search on a line by byzantine robots. In ISAAC, pages 27:1–27:12, 2016.
  • [23] J. Czyzowicz, K. Georgiou, E. Kranakis, L. Narayanan, J. Opatrny, and B. Vogtenhuber. Evacuating robots from a disc using face to face communication. In CIAC 2015, pages 140–152, 2015.
  • [24] J. Czyzowicz, E. Kranakis, D. Krizanc, L. Narayanan, and Opatrny J. Search on a line with faulty robots. In PODC, pages 405–414. ACM, 2016.
  • [25] J. Czyzowicz, E. Kranakis, D. Krizanc, L. Narayanan, J. Opatrny, and M. Shende. Linear search with terrain-dependent speeds. In CIAC, pages 430–441, 2017.
  • [26] J. Czyzowicz, E. Kranakis, K. Krizanc, L. Narayanan, J. Opatrny, and S. Shende. Wireless autonomous robot evacuation from equilateral triangles and squares. In ADHOCNOW, pages 181–194. Springer, 2015.
  • [27] E. D. Demaine, S. P. Fekete, and S. Gal. Online searching with turn cost. Theoretical Computer Science, 361(2):342–355, 2006.
  • [28] Y. Emek, T. Langner, D. Stolz, J. Uitto, and R. Wattenhofer. How many ants does it take to find the food? Theor. Comput. Sci., page 608:255–267, 2015.
  • [29] S. Gal. Search Games. Wiley Encyclopedia for Operations Research and Management Science, 2011.
  • [30] K. Georgiou, G. Karakostas, and E. Kranakis. Search-and-fetch with one robot on a disk - (track: Wireless and geometry). In ALGOSENSORS, pages 80–94, 2016.
  • [31] K. Georgiou, G. Karakostas, and E. Kranakis. Search-and-fetch with 2 robots on a disk - wireless and face-to-face communication models. In ICORES, pages 15–26. SciTePress, 2017.
  • [32] F. Hoffmann, C. Icking, R. Klein, and K. Kriegel. The polygon exploration problem. SIAM Journal on Computing, 31(2):577–600, 2001.
  • [33] M.-Y. Kao, J. H. Reif, and S. R. Tate. Searching in an unknown environment: An optimal randomized algorithm for the cow-path problem. Information and Computation, 131(1):63–79, 1996.
  • [34] J. Kleinberg. On-line search in a simple polygon. In SODA, pages 8–15. SIAM, 1994.
  • [35] I. Lamprou, R. Martin, and S. Schewe. Fast two-robot disk evacuation with wireless communication. In DISC, pages 1–15, 2016.
  • [36] D. Pattanayak, H. Ramesh, P.S. Mandal, and S. Schmid. Evacuating two robots from two unknown exits on the perimeter of a disk with wireless communication. In ICDCN, pages 20:1–20:4, 2018.
  • [37] O. Reingold. Undirected st-connectivity in log-space. In STOC, pages 376–385, 2005.
  • [38] H.-A. Rollik. Automaten in planaren graphen. Acta Informatica, 13(3):287– 298, 1980.

Appendix A Figures

Refer to caption
Figure 1: The competitive ratio of algorithm 𝒩s,r,k\mathcal{N}_{s,r,k} (vertical axis) for the entire spectrum of cb9cb\geq 9 (horizontal axis). Red curve corresponds to the case cbγ1cb\leq\gamma_{1}, blue curve to the case cb(γ1,γ2)cb\in(\gamma_{1},\gamma_{2}) and green curve to the case cbγ2cb\geq\gamma_{2}. The curve is continuous and differentiable for all cb9cb\geq 9.
Refer to caption
Figure 2: Comparison between the competitive ratio achieved by using optimal speed parameters to NLPcb\textsc{NLP}_{c}^{b} of Theorem 3.2 (calculated numerically using software) and the competitive ratio achieved by the choices of Theorem 3.4. The vertical axis is the difference of the competitive ratios, and the horizontal axis corresponds to the values of cb(γ1,γ2)cb\in(\gamma_{1},\gamma_{2}) (for all other values of cbcb the difference is provably 0).
Refer to caption
Figure 3: A representation of position (x-axis, vertical dashed line is 0) and time (y-axis), and the trajectory followed by the two robots (solid lines). The two diagonal dashed lines form the ``1/3 cone'' of Theorem 4.1.
Refer to caption
Figure 4: The robots' behavior when the exit is found by LL is indicated by the bold line. In the first case (left), the catch-up speed is slower than 11 (and the rendezvous is realized at the turning point of the non-finder), whereas it is 11 in the second case (right).

Appendix B Observation B.1

Observation B.1.

EEcb\textsc{EE}_{c}^{b} is well defined for each b,c>0b,c>0 with bc1bc\geq 1, and the optimal solution, given that instance dd is known, equals 2dc2\frac{2d}{c^{2}}.

Proof: [Proof of Observation B.1] Given that the location of the exit is known, and by symmetry, it is immediate that both robots have the same optimal speed, call it ss, and they move in the direction of the exit. The induced evacuation time is then d/sd/s, and the induced evacuation energy is 2ds22d\cdot s^{2}. For a feasible solution we require that d/scdd/s\leq c\cdot d and that 0<sb0<s\leq b, and hence, the optimal offline solution is obtained as the solution to mins{s2:1/csb}.\min_{s}\{s^{2}:~1/c\leq s\leq b\}. For a feasible solution to exist, we need bc1bc\geq 1. Moreover, it is immediate that the optimal choice is s=1/cs=1/c, inducing energy consumption 2ds2=2d/c22d\cdot s^{2}=2d/c^{2}. ∎

Appendix C Proofs Omitted from Section 3.

C.1 Lemma 3.1

Proof: [Proof of Lemma 3.1] Consider the moment that the exit is located, after time d/sd/s time of searching. The robot that now chases the other speed-ss robot at constant speed r>sr>s will reach it after 2d/(rs)2d/(r-s) time. To see this note that the configuration is equivalent to that the speed-ss robot is immobile and the other robot moves at speed rsr-s, having to traverse a total distance of 2d2d. Moreover, the speed-ss robot traverses an additional length 2ds/(rs)2ds/(r-s) segment till it is caught, being a total of 2ds/(rs)+2d2ds/(r-s)+2d away from the exit. Once robots meet, the walk to the exit at speed kk, which takes additional time (2ds/(rs)+2d)/k(2ds/(r-s)+2d)/k. Overall the evacuation time equals

ds+2drs+2ds/(rs)+2dk=d(2(k+r)k(rs)+1s).\frac{d}{s}+\frac{2d}{r-s}+\frac{2ds/(r-s)+2d}{k}=d\left(\frac{2(k+r)}{k(r-s)}+\frac{1}{s}\right).

Similarly we compute the total energy till both robots reach the exit. The energy spent by the finder is

ds2+(2dsrs+2d)r2+(2dsrs+2d)k2,d\cdot s^{2}+\left(\frac{2ds}{r-s}+2d\right)\cdot r^{2}+\left(\frac{2ds}{r-s}+2d\right)\cdot k^{2},

while the energy spent by the non finder is

(d+2dsrs)s2+(2dsrs+2d)k2.\left(d+\frac{2ds}{r-s}\right)\cdot s^{2}+\left(\frac{2ds}{r-s}+2d\right)\cdot k^{2}.

Adding the two quantities and simplifying gives the promised formula. ∎

C.2 Theorem 3.2

Proof: [Proof of Theorem 3.2] Note that in NLPc\textsc{NLP}_{c}^{\infty} that aims to provide a solution to EEc\textsc{EE}_{c}, constraints s,r,ks,r,k\leq\infty are simply omitted. In particular, the theorem above claims that when b=b=\infty, i.e. when speeds are unbounded, algorithm 𝒩s,r,k\mathcal{N}_{s,r,k} always admits some feasible solution. In what follows, we prove all claims of the theorem.

By Lemma 3.1, the energy performance of 𝒩s,r,k\mathcal{N}_{s,r,k} equals 2dE(s,r,k)2d\cdot E\left({s,r,k}\right), and the induced evacuation time is dT(s,r,k)d\cdot T\left({s,r,k}\right). For the values of s,r,ks,r,k to be feasible, we need that 0<s,r,kb0<s,r,k\leq b, that r>kr>k and that dT(s,r,k)cdd\cdot T\left({s,r,k}\right)\leq cd. Clearly the latter time constraint simplifies to the time constraint of NLPcb\textsc{NLP}_{c}^{b}, while the objective value can be scaled by d>0d>0 without affecting the optimizers to the NLP, if such optimizers exist. Finally note that even though the strict inequalities become non strict inequalities in the NLP, speeds evaluations for which any of s,r,ks,r,k is 0 or r=kr=k violates the time constraint (for any fixed c>0c>0). Therefore, NLPcb\textsc{NLP}_{c}^{b} correctly formulates the problem of choosing optimal values for 𝒩s,r,k\mathcal{N}_{s,r,k} for solving EEcb\textsc{EE}_{c}^{b}.

The next two lemmas show that the Naive algorithm can solve problem EEcb\textsc{EE}_{c}^{b} for the entire spectrum of c,bc,b values for which the problem admits solutions, as per Lemma 2.1.

Lemma C.1.

For every cc, problem NLPc\textsc{NLP}_{c}^{\infty} admits an optimal solution.

Proof: [Proof of Lemma C.1] Consider the redundant constraints s,r,k1/cs,r,k\geq 1/c that can be derived by the existing constraints of NLPc\textsc{NLP}_{c}^{\infty} (note that if all speeds are not at least 1/c1/c then clearly the time constraint is violated). For the same reason, it is also easy to see that rs1/cr-s\geq 1/c, since again we would have a violation of the time constraint.

Next, it is easy to check that s=7/c,r=14/c,k=7/cs=7/c,r=14/c,k=7/c is a feasible solution, hence the NLP is not infeasible. The value of the objective for this evaluation is 686/c2686/c^{2}. But then, notice that the objective is bounded from below by s2+r2+2k2s^{2}+r^{2}+2k^{2}. Hence, if an optimal solution exists, constraints s,r,k686/cs,r,k\leq\sqrt{686}/c are valid for the optimizers. We may add these constraints to NLPc\textsc{NLP}_{c}^{\infty}, resulting into a compact (closed and bounded) feasible region. But then, note that the objective is continuous over the new compact feasible region, hence from the extreme value theorem it attains a minimum. ∎

Lemma C.2.

There exist s,r,ks,r,k for which 𝒩s,r,k\mathcal{N}_{s,r,k} induces a feasible solution to EEcb\textsc{EE}_{c}^{b} if and only if c9/bc\geq 9/b.

Proof: [Proof of Lemma C.2] Consider the problem of minimizing completion time of the Naive Algorithm, given that the speeds are all bounded above by bb. The corresponding NLP that solves the problem reads as.

min\displaystyle\min 2(k+r)k(rs)+1s\displaystyle\frac{2(k+r)}{k(r-s)}+\frac{1}{s} (2)
s.t.\displaystyle s.t.~~ rs\displaystyle r\geq s
0s,r,kb\displaystyle 0\leq s,r,k\leq b

Note that it is enough to prove that the optimal value to (2) is 9/b9/b. Indeed, that would imply that no speeds exist that induce completion time less than 9/b9/b, making the corresponding feasible region of NLPcb\textsc{NLP}_{c}^{b} empty if c<9/bc<9/b.

Now we show that the optimal value to (2) is 9/b9/b, by showing that the unique optimizers to the NLP are r=k=br=k=b and s=b/3s=b/3. Indeed, note that

r(2(k+r)k(rs)+1s)=2(k+s)k(rs)2\frac{\partial}{\partial r}\left(\frac{2(k+r)}{k(r-s)}+\frac{1}{s}\right)=-\frac{2(k+s)}{k(r-s)^{2}}

which is strictly negative for all feasible s,r,ks,r,k with r>sr>s. Hence, there is no optimal solution for which r<br<b, as otherwise by increasing rr one could improve the value of the objective. Similarly we observe that

k(2(k+r)k(rs)+1s)=2rk2(rs)\frac{\partial}{\partial k}\left(\frac{2(k+r)}{k(r-s)}+\frac{1}{s}\right)=-\frac{2r}{k^{2}(r-s)}

which is again strictly negative for all feasible s,r,ks,r,k with r>sr>s. Hence, there is no optimal solution for which k<bk<b, as otherwise by increasing kk one could improve the value of the objective.

To conclude, in an optimal solution to (2) we have that r=k=br=k=b, and hence one needs to find ss minimizing g(s,b,b)=4bs+1sg(s,b,b)=\frac{4}{b-s}+\frac{1}{s}. For this we compute

sg(s,b,b)=4(bs)21s2\frac{\partial}{\partial s}g(s,b,b)=\frac{4}{(b-s)^{2}}-\frac{1}{s^{2}}

and it is easy to see that g(s,b,b)=0g(s,b,b)=0 if and only if s=b/3s=b/3 or s=bs=-b (and the latter is infeasible). At the same time, g(s,b,b)g(s,b,b) is convex when sbs\leq b because 2s2g(s,b,b)=8(bs)3+2s3>0\frac{\partial^{2}}{\partial s^{2}}g(s,b,b)=\frac{8}{(b-s)^{3}}+\frac{2}{s^{3}}>0, hence s=b/3s=b/3 corresponds to the unique minimizer. ∎

The last component of Theorem 3.2 that requires justification pertains to the competitive ratio. Now fix b,c>0b,c>0 for which bc9bc\geq 9, and let E(s0,r0,k0)E\left({s_{0},r_{0},k_{0}}\right) be the optimal solution to NLPcb\textsc{NLP}_{c}^{b} (corresponding to the optimal choices of algorithm 𝒩s,r,k\mathcal{N}_{s,r,k}). By Lemma 3.1 the induced energy consumption is 2dE(s0,r0,k0)2d\cdot E\left({s_{0},r_{0},k_{0}}\right). Then, the competitive ratio of the algorithm is supd>0c22d2dE(s0,r0,k0)=c2E(s0,r0,k0).\sup_{d>0}\frac{c^{2}}{2d}~2d\cdot E\left({s_{0},r_{0},k_{0}}\right)=c^{2}\cdot E\left({s_{0},r_{0},k_{0}}\right).

C.3 Theorem 3.3

Proof: [Proof of Theorem 3.3] First we observe that s=σc,r=ρc,k=κcs=\frac{\sigma}{c},r=\frac{\rho}{c},k=\frac{\kappa}{c} are indeed feasible to NLPc\textsc{NLP}_{c}^{\infty} (for every c>0c>0), since

T(σc,ρc,κc)=c(2(κ+ρ)κ(ρσ)+1σ)T\left({\frac{\sigma}{c},\frac{\rho}{c},\frac{\kappa}{c}}\right)=c\left(\frac{2(\kappa+\rho)}{\kappa(\rho-\sigma)}+\frac{1}{\sigma}\right)

and in particular, for the values of σ,ρ,κ\sigma,\rho,\kappa described above we have (2(κ+ρ)κ(ρσ)+1σ)1\left(\frac{2(\kappa+\rho)}{\kappa(\rho-\sigma)}+\frac{1}{\sigma}\right)\approx 1 (from the formal definition of σ,ρ,κ\sigma,\rho,\kappa that appears later, it will be clear that expression will be exactly equal to 1). Moreover, by Theorem 3.2, the competitive ratio of 𝒩σc,ρc,κc\mathcal{N}_{\frac{\sigma}{c},\frac{\rho}{c},\frac{\kappa}{c}} is

c2E(σc,ρc,κc)=ρ(2κ2+ρ2+σ2)ρσ,c^{2}\cdot E\left({\frac{\sigma}{c},\frac{\rho}{c},\frac{\kappa}{c}}\right)=\frac{\rho\left(2\kappa^{2}+\rho^{2}+\sigma^{2}\right)}{\rho-\sigma},

as claimed.

In the remaining of the section we prove that the choices for s,r,ks,r,k of Theorem 3.3 are indeed optimal for 𝒩s,r,k\mathcal{N}_{s,r,k}. First we establish a structural property of optimal speeds choices for 𝒩s,r,k\mathcal{N}_{s,r,k}.

Lemma C.3.

For any c>0c>0, optimal solutions to NLPc\textsc{NLP}_{c}^{\infty} satisfy constraint T(s,r,k)cT\left({s,r,k}\right)\leq c tightly.

Proof: Consider an optimal solution s¯,r¯,k¯\bar{s},\bar{r},\bar{k}. As noted before, we must have s¯,r¯,k¯>0\bar{s},\bar{r},\bar{k}>0 and r¯>s¯\bar{r}>\bar{s}, as otherwise the values would be infeasible.

Next note that the time constraint can be rewritten as

k2rscrscs2rsk\geq\frac{2rs}{crs-cs^{2}-r-s}

For the sake of contradiction, assume that the time constraint is not tight for s¯,r¯,k¯\bar{s},\bar{r},\bar{k}. Then, there is ϵ>0\epsilon>0 so that s¯,r¯,k\bar{s},\bar{r},k^{\prime} is a feasible solution, where k=k¯ϵ>0k^{\prime}=\bar{k}-\epsilon>0. But then, the objective value strictly decreases, a contradiction to optimality. ∎

We will soon derive the optimizers to NLPc\textsc{NLP}_{c}^{\infty} using Karush-Kuhn-Tucker (KKT) conditions. Before that, we observe that solutions are scalable with respect to cc, which will also allow us to simplify our calculations.

Lemma C.4.

Let s,r,ks^{\prime},r^{\prime},k^{\prime} be the optimizers to NLP1\textsc{NLP}_{1}^{\infty} inducing optimal energy EE. Then, for any cc, the optimizers to NLPc\textsc{NLP}_{c}^{\infty} are s¯=s/c,r¯=r/c,k¯=k/c\bar{s}=s^{\prime}/c,\bar{r}=r^{\prime}/c,\bar{k}=k^{\prime}/c, and the induced optimal energy is 1c2E\frac{1}{c^{2}}E.

Proof: Note that the triplet (s,r,k)(s,r,k) is feasible to NLPc\textsc{NLP}_{c}^{\infty} (for a specific cc) if and only if the triplet (cs,cr,ck)(c\cdot s,c\cdot r,c\cdot k) is feasible to NLP1\textsc{NLP}_{1}^{\infty}. Moreover, it is straightforward that when speeds are scaled by cc, the induced energy is scaled by c2c^{2}. Hence, for every c>0c>0 there is a bijection between feasible (and optimal) solutions to  NLPc\textsc{NLP}_{c}^{\infty} and  NLP1\textsc{NLP}_{1}^{\infty}. ∎

We are therefore motivated to solve NLP1\textsc{NLP}_{1}^{\infty}, and that will allow us to derive the optimizers for NLPc\textsc{NLP}_{c}^{\infty}, for any c>0c>0.

Lemma C.5.

The optimal solution to NLP1\textsc{NLP}_{1}^{\infty} is obtained for

s=σ2.65976,r=ρ11.3414,k=κ6.63709s=\sigma\approx 2.65976,~r=\rho\approx 11.3414,~k=\kappa\approx 6.63709

and the optimal NLP value is ρ(2κ2+ρ2+σ2)ρσ292.37\frac{\rho\left(2\kappa^{2}+\rho^{2}+\sigma^{2}\right)}{\rho-\sigma}\approx 292.37.

Proof: By KKT conditions, we know that, necessarily, all minimizers of E(s,r,k)E\left({s,r,k}\right) satisfy the condition that E(s,r,k)-\nabla E\left({s,r,k}\right) is a conical combination of tight constraints (for the optimizers). Lemma C.3 asserts that T(s,r,k)=1T\left({s,r,k}\right)=1 has to be satisfied for all optimizers s,r,ks,r,k. At the same time, recall that, by the proof of Lemma C.1, none of the constraints rsr\geq s and s,r,k0s,r,k\geq 0 can be tight for an optimizer. Hence, KKT conditions imply that any optimizer s,r,ks,r,k satisfies, necessarily, the following system of nonlinear constraints

E(s,r,k)\displaystyle-\nabla E\left({s,r,k}\right) =λT(s,r,k)\displaystyle=\lambda\nabla T\left({s,r,k}\right)
T(s,r,k)\displaystyle T\left({s,r,k}\right) =1\displaystyle=1
λ\displaystyle\lambda 0\displaystyle\geq 0

More explicitly, the first equality constraints is

(r2r(k2+r2)(rs)22k2s2r3+3r2s+s3(rs)24krsr)\displaystyle\left(\begin{array}[]{c}r-\frac{2r\left(k^{2}+r^{2}\right)}{(r-s)^{2}}\\ \frac{2k^{2}s-2r^{3}+3r^{2}s+s^{3}}{(r-s)^{2}}\\ \frac{4kr}{s-r}\end{array}\right) =λ(2(k+r)k(rs)21s22(k+s)k(rs)22rk2(sr))\displaystyle=\lambda\left(\begin{array}[]{c}\frac{2(k+r)}{k(r-s)^{2}}-\frac{1}{s^{2}}\\ -\frac{2(k+s)}{k(r-s)^{2}}\\ \frac{2r}{k^{2}(s-r)}\end{array}\right)

From the 3rd coordinates of the gradients, we obtain that λ=2k3\lambda=2k^{3}, which directly implies that the dual multiplier λ\lambda preserves the correct sign for the necessary optimality conditions.

Hence, the original system of nonlinear constraints is equivalent to that

r2r(k2+r2)(rs)2\displaystyle r-\frac{2r\left(k^{2}+r^{2}\right)}{(r-s)^{2}} =2k3(2(k+r)k(rs)21s2)\displaystyle=2k^{3}\left(\frac{2(k+r)}{k(r-s)^{2}}-\frac{1}{s^{2}}\right)
2k2s+2r33r2ss3\displaystyle-2k^{2}s+2r^{3}-3r^{2}s-s^{3} =4k2(k+s)\displaystyle=4k^{2}(k+s)
2(k+r)k(rs)+1s\displaystyle\frac{2(k+r)}{k(r-s)}+\frac{1}{s} =1\displaystyle=1

Using software numerical methods, we see that the above algebraic system admits the following 3 real roots for (s,r,k)(s,r,k):

(2.659764883844293,11.341425445393606,6.637089776204052)\displaystyle(2.659764883844293,~11.341425445393606,~6.637089776204052) (multiplicity 1)\displaystyle(\textrm{multiplicity 1})
(0.6115006613361799,0.47813267995355124,1.0972211311317337)\displaystyle(-0.6115006613361799,~0.47813267995355124,~1.0972211311317337) (multiplicity 2)\displaystyle(\textrm{multiplicity 2})

Since also all speeds are nonnegative, we obtain the unique candidate optimizer

(σ,ρ,κ)=(2.65976,11.3414,6.63709).(\sigma,\rho,\kappa)=(2.65976,11.3414,6.63709).

To verify that indeed (σ,ρ,κ)(\sigma,\rho,\kappa) is a minimizer, we compute

2E(s,r,k)=(4r(k2+r2)(rs)3(r+s)(2k2+r2+s24rs)(rs)34kr(rs)2(r+s)(2k2+r2+s24rs)(rs)32(r33sr2+3s2r+s3+2k2s)(rs)34ks(rs)24kr(rs)24ks(rs)24rrs).\nabla^{2}E\left({s,r,k}\right)=\left(\begin{array}[]{ccc}\frac{4r\left(k^{2}+r^{2}\right)}{(r-s)^{3}}&\frac{(r+s)\left(-2k^{2}+r^{2}+s^{2}-4rs\right)}{(r-s)^{3}}&\frac{4kr}{(r-s)^{2}}\\ \frac{(r+s)\left(-2k^{2}+r^{2}+s^{2}-4rs\right)}{(r-s)^{3}}&\frac{2\left(r^{3}-3sr^{2}+3s^{2}r+s^{3}+2k^{2}s\right)}{(r-s)^{3}}&-\frac{4ks}{(r-s)^{2}}\\ \frac{4kr}{(r-s)^{2}}&-\frac{4ks}{(r-s)^{2}}&\frac{4r}{r-s}\\ \end{array}\right).

Moreover,

2E(σ,ρ,κ)=(11.97181.563333.994851.563332.831250.9368643.994850.9368645.22546)\nabla^{2}E\left({\sigma,\rho,\kappa}\right)=\left(\begin{array}[]{ccc}11.9718&-1.56333&3.99485\\ -1.56333&2.83125&-0.936864\\ 3.99485&-0.936864&5.22546\\ \end{array}\right)

which has eigenvalues 14.1183,3.41098,2.4992714.1183,3.41098,2.49927, hence it is PSD. As a result, f(s,r,k)f(s,r,k) is locally convex at (σ,ρ,κ)(\sigma,\rho,\kappa), and therefore (σ,ρ,κ)(\sigma,\rho,\kappa) is a local minimizer to NLP1\textsc{NLP}_{1}^{\infty}. As we showed earlier, (σ,ρ,κ)(\sigma,\rho,\kappa) is the only candidate optimizer, hence a global minimizer as well.

Lemma C.5 together with Lemma C.4 imply that for any c>0c>0 the optimal solution to NLPc\textsc{NLP}_{c}^{\infty} is exactly for

s=σc,r=ρc,k=κcs=\frac{\sigma}{c},~r=\frac{\rho}{c},~k=\frac{\kappa}{c}

and hence, the proof of Theorem 3.3 follows. ∎

C.4 Lemma 3.6

Proof: [Proof of Lemma 3.6]

An immediate corollary from the proof of Lemma C.2 (within the proof of Theorem 3.2) is the following

Corollary C.6.

The unique solution to NLPcb\textsc{NLP}_{c}^{b} when c=9/bc=9/b is given by

r=k=b,s=b3,r=k=b,s=\frac{b}{3},

inducing energy 28b2d3\frac{28b^{2}d}{3}, and competitive ratio 14(cb)23=378\frac{14(cb)^{2}}{3}=378.

Next we find solutions for c>9/bc>9/b so that r,kbr,k\leq b remain tight. Since, when c=9/bc=9/b, there is only one optimizer s=b/3,r=b,k=bs=b/3,r=b,k=b, two inequality constraints are tight. The next calculations investigate the spectrum of cc for which the same constraints remain tight for the optimizer.

We write 1st order necessary optimality conditions for NLPcb\textsc{NLP}_{c}^{b}, given that the candidate optimizer satisfies the time constraint, and the two r,kbr,k\leq b speed bound constraints tightly

E(s,r,k)\displaystyle-\nabla E\left({s,r,k}\right) =λ1T(s,r,k)+λ2(010)+λ3(001)\displaystyle=\lambda_{1}\nabla T\left({s,r,k}\right)+\lambda_{2}\left(\begin{array}[]{c}0\\ 1\\ 0\end{array}\right)+\lambda_{3}\left(\begin{array}[]{c}0\\ 0\\ 1\end{array}\right)
T(s,r,k)\displaystyle T\left({s,r,k}\right) =c\displaystyle=c
r\displaystyle r =b\displaystyle=b
k\displaystyle k =b\displaystyle=b
λ1,λ2,λ3\displaystyle\lambda_{1},\lambda_{2},\lambda_{3} 0\displaystyle\geq 0

From the tight time constraint, and solving for ss we obtain that

s1,2=±b2c210bc+9+bc32cs_{1,2}=\frac{\pm\sqrt{b^{2}c^{2}-10bc+9}+bc-3}{2c}

For each s{s1,s2}s\in\{s_{1},s_{2}\}, the first gradient equality defines a linear system over λ1,λ2,λ3\lambda_{1},\lambda_{2},\lambda_{3} whose solutions are

λ1=(s3)s23s1,λ2=5s39s+2(s1)(3s1),λ3=2(s33s26s+2)(s1)(3s1).\lambda_{1}=\frac{(s-3)s^{2}}{3s-1},~~\lambda_{2}=-\frac{-5s^{3}-9s+2}{(s-1)(3s-1)},~~\lambda_{3}=-\frac{2\left(s^{3}-3s^{2}-6s+2\right)}{(s-1)(3s-1)}.
λ1=bs2(3bs)b3s,λ2=2b39b2s5s3(b3s)(bs),λ3=2(2b36b2s3bs2+s3)(b3s)(bs)\lambda_{1}=\frac{bs^{2}(3b-s)}{b-3s},~~\lambda_{2}=-\frac{2b^{3}-9b^{2}s-5s^{3}}{(b-3s)(b-s)},~~\lambda_{3}=-\frac{2\left(2b^{3}-6b^{2}s-3bs^{2}+s^{3}\right)}{(b-3s)(b-s)}

respectively. As long as all dual multiplies λi=λi(s)\lambda_{i}=\lambda_{i}(s) are positive, corresponding solution (s,b,b)(s,b,b) is optimal to RcR_{c}^{\prime}, provided that 2f(s,b,b)0\nabla^{2}f(s,b,b)\succ 0 .

First we claim that s1s_{1} cannot be part of an optimizer. Indeed,

λ1(s1)=b(5bc(bc9)(bc1)+3)(bc+(bc9)(bc1)3)24c2(bc+3(bc9)(bc1)9)\lambda_{1}(s_{1})=-\frac{b\left(5bc-\sqrt{(bc-9)(bc-1)}+3\right)\left(bc+\sqrt{(bc-9)(bc-1)}-3\right)^{2}}{4c^{2}\left(bc+3\sqrt{(bc-9)(bc-1)}-9\right)}

Recall that bc>9bc>9, and hence the denominator of λ1(s1)\lambda_{1}(s_{1}) as well as bc+(bc9)(bc1)3bc+\sqrt{(bc-9)(bc-1)}-3 are strictly positive. But then, the sign of λ1(s1)\lambda_{1}(s_{1}) is exactly the opposite of 5bc(bc9)(bc1)+35bc-\sqrt{(bc-9)(bc-1)}+3. Define function h(x):=5x(x9)(x1)+3h(x):=5x-\sqrt{(x-9)(x-1)}+3 over the domain x>9x>9. It is easy to verify that h(x)h(x) preserves positive sign (in fact minx9h(x)=h(53(6+3))=86+28>0\min_{x\geq 9}h(x)=h\left(\frac{5}{3}\left(\sqrt{6}+3\right)\right)=8\sqrt{6}+28>0 Hence, λ1(s1)<0\lambda_{1}(s_{1})<0 that concludes our claim.

Next we investigate the spectrum of cc for which all λi(s2)\lambda_{i}(s_{2}) remain non-negative.

Our next claim is that for all bc>9bc>9 we have that s2(c)<b/3s_{2}(c)<b/3. Indeed, consider function

d(x):=3x210x+9bx+9.d(x):=3\sqrt{x^{2}-10x+9}-bx+9.

It is easy to see that d(bc)=6c(b3s2(c))d(bc)=6c\left(\frac{b}{3}-s_{2}(c)\right). But then, elementary calculations show that minx9d(x)=d(9)=0\min_{x\geq 9}d(x)=d(9)=0, proving that s2(c)<b/3s_{2}(c)<b/3 as claimed.

Next we investigate the sign of λ1(s2),λ2(s2),λ3(s2)\lambda_{1}(s_{2}),\lambda_{2}(s_{2}),\lambda_{3}(s_{2}). For this, introduce function t(x)=(x9)(x1)t(x)=\sqrt{(x-9)(x-1)}, and note that

λ1(s2)\displaystyle\lambda_{1}(s_{2}) =b(bc+t(bc)+3)2(5bc+t(bc)+3)4c2(bc3(t(bc)+3)),\displaystyle=-\frac{b\left(-bc+t(bc)+3\right)^{2}\left(5bc+t(bc)+3\right)}{4c^{2}\left(bc-3\left(t(bc)+3\right)\right)},
λ2(s2)\displaystyle\lambda_{2}(s_{2}) =30(t(bc)+3)+bc(bc(3bc+3t(bc)+32)5(5t(bc)+11))4c(bc9),\displaystyle=\frac{30\left(t(bc)+3\right)+bc\left(bc\left(-3bc+3t(bc)+32\right)-5\left(5t(bc)+11\right)\right)}{4c(bc-9)},
λ3(s2)\displaystyle\lambda_{3}(s_{2}) =bc(23t(bc)+bc(3bc+3t(bc)+22)+49)12(t(bc)+3)4c(bc9).\displaystyle=\frac{bc\left(-23t(bc)+bc\left(-3bc+3t(bc)+22\right)+49\right)-12\left(t(bc)+3\right)}{4c(bc-9)}.

Claim 1: λ1(s2)>0\lambda_{1}(s_{2})>0 for all c>9/bc>9/b.
Define d1(x)=x3(3+t(x))d_{1}(x)=x-3(3+t(x)) and d2(x)=3+5x+t(x)d_{2}(x)=3+5x+t(x). Note that sign(λ1(s2))=sign(d1(bc))sign(d2(bc))sign\left(\lambda_{1}(s_{2})\right)=-sign(d_{1}(bc))\cdot sign(d_{2}(bc)). Simple calculus shows that d1(x)d_{1}(x) is strictly decreasing in x9x\geq 9, and d1(9)=0d_{1}(9)=0, and therefore d1(bc)<0d_{1}(bc)<0 for all c>9/bc>9/b. Similarly, it is easy to see that d2(x)d_{2}(x) is strictly increasing in x>9x>9, and d2(9)=45d_{2}(9)=45. Therefore d2(bc)>0d_{2}(bc)>0 for all c>9/bc>9/b. Overall this implies that λ1(s2)\lambda_{1}(s_{2}) is positive for all c>9/bc>9/b.

Claim 2: λ2(s2)>0\lambda_{2}(s_{2})>0 for all c(9/b,9.72307/b)c\in(9/b,9.72307/b).
First we observe that the denominator of λ2(s2)\lambda_{2}(s_{2}) preserves positive sign for c>9/bc>9/b. So we focus on the sign of the numerator we we abbreviate by d3(x)=30(3+t(x))+x(x(323x+3t(x))5(11+5t(x)))d_{3}(x)=30(3+t(x))+x(x(32-3x+3t(x))-5(11+5t(x))). Note that d3(x)=0d_{3}(x)=0 is equivalent to that

(3x225x+30)t(x)(3x332x2+55x90)=0\displaystyle\left(3x^{2}-25x+30\right)t(x)-\left(3x^{3}-32x^{2}+55x-90\right)=0
\displaystyle\Leftrightarrow (3x225x+30)2t2(x)(3x332x2+55x90)2=0\displaystyle\left(3x^{2}-25x+30\right)^{2}t^{2}(x)-\left(3x^{3}-32x^{2}+55x-90\right)^{2}=0
\displaystyle\Leftrightarrow 8(x9)x(3x(x(2x25)+60)175)=0\displaystyle-8(x-9)x(3x(x(2x-25)+60)-175)=0

Degree-3 polynomial 3x(x(2x25)+60)1753x(x(2x-25)+60)-175 has only one real root, which is

118(35(19210+1055)3+14242525920103+75)9.72307\frac{1}{18}\left(3\sqrt[3]{5\left(192\sqrt{10}+1055\right)}+\sqrt[3]{142425-25920\sqrt{10}}+75\right)\approx 9.72307

Hence, λ2(s2)>0\lambda_{2}(s_{2})>0 for all c(9/b,9.72307/b)c\in(9/b,9.72307/b)

Claim 3: λ3(s2)>0\lambda_{3}(s_{2})>0 for all c(9/b,9.06609/b)c\in(9/b,9.06609/b).
First we observe that the denominator of λ3(s2)\lambda_{3}(s_{2}) preserves positive sign for c>9/bc>9/b. So we focus on the sign of the numerator we we abbreviate by d4(x)=x(x(3t(x)3x+22)23t(x)+49)12(t(x)+3)d_{4}(x)=x(x(3t(x)-3x+22)-23t(x)+49)-12(t(x)+3). Note that d4(x)=0d_{4}(x)=0 is equivalent to that

(3x223x12)t(x)(3x322x249x+36)=0\displaystyle\left(3x^{2}-23x-12\right)t(x)-\left(3x^{3}-22x^{2}-49x+36\right)=0
\displaystyle\Leftrightarrow (3x223x12)2t2(x)(3x322x249x+36)2=0\displaystyle\left(3x^{2}-23x-12\right)^{2}t^{2}(x)-\left(3x^{3}-22x^{2}-49x+36\right)^{2}=0
\displaystyle\Leftrightarrow 16(x9)x(3x(2(x9)x3)+49)=0\displaystyle-16(x-9)x(3x(2(x-9)x-3)+49)=0

The roots of degree-3 polynomial 3x(2(x9)x3)+493x(2(x-9)x-3)+49 are

γ1\displaystyle\gamma_{1} =3+38cos(13tan1(1271512))9.06609\displaystyle=3+\sqrt{38}\cos\left(\frac{1}{3}\tan^{-1}\left(\frac{127}{151\sqrt{2}}\right)\right)\approx 9.06609
γ\displaystyle\gamma^{\prime} =3+572sin(13tan1(1271512))+192(cos(13tan1(1271512)))0.916629\displaystyle=3+\sqrt{\frac{57}{2}}\sin\left(\frac{1}{3}\tan^{-1}\left(\frac{127}{151\sqrt{2}}\right)\right)+\sqrt{\frac{19}{2}}\left(-\cos\left(\frac{1}{3}\tan^{-1}\left(\frac{127}{151\sqrt{2}}\right)\right)\right)\approx 0.916629
γ′′\displaystyle\gamma^{\prime\prime} =3572sin(13tan1(1271512))+192(cos(13tan1(1271512)))0.982723\displaystyle=3-\sqrt{\frac{57}{2}}\sin\left(\frac{1}{3}\tan^{-1}\left(\frac{127}{151\sqrt{2}}\right)\right)+\sqrt{\frac{19}{2}}\left(-\cos\left(\frac{1}{3}\tan^{-1}\left(\frac{127}{151\sqrt{2}}\right)\right)\right)\approx-0.982723

We conclude that λ3(s2)\lambda_{3}(s_{2}) preserves positive sign for all c(9/b,γ1/b)c\in(9/b,\gamma_{1}/b).

Overall, we have shown that feasible solution s0=b2c210bc+9+bc32c,r0=k0=bs_{0}=\frac{-\sqrt{b^{2}c^{2}-10bc+9}+bc-3}{2c},r_{0}=k_{0}=b satisfies necessary 1st order optimality conditions. We proceed by checking that s0,r0,k0s_{0},r_{0},k_{0} satisfy 2nd order sufficient conditions, which amounts to showing that 2f(s0,b,b)0\nabla^{2}f(s_{0},b,b)\succ 0. Indeed,

2f(s0,b,b)=b3(bs0)3(8(b+s0)(b2+4s0bs02)b344s0b(b+s0)(b2+4s0bs02)b32(b3s0b2+3s02b+s03)b34s0(s0b)b244s0b4s0(s0b)b24(bs0)2b2)\nabla^{2}f(s_{0},b,b)=\frac{b^{3}}{(b-s_{0})^{3}}\left(\begin{array}[]{ccc}8&-\frac{\left(b+s_{0}\right)\left(b^{2}+4s_{0}b-s_{0}^{2}\right)}{b^{3}}&4-\frac{4s_{0}}{b}\\ -\frac{\left(b+s_{0}\right)\left(b^{2}+4s_{0}b-s_{0}^{2}\right)}{b^{3}}&\frac{2\left(b^{3}-s_{0}b^{2}+3s_{0}^{2}b+s_{0}^{3}\right)}{b^{3}}&\frac{4s_{0}\left(s_{0}-b\right)}{b^{2}}\\ 4-\frac{4s_{0}}{b}&\frac{4s_{0}\left(s_{0}-b\right)}{b^{2}}&\frac{4\left(b-s_{0}\right){}^{2}}{b^{2}}\\ \end{array}\right)

By setting q:=s0/b=b2c210bc+9+bc32bcq:=s_{0}/b=\frac{-\sqrt{b^{2}c^{2}-10bc+9}+bc-3}{2bc}, we obtain the simpler form

2f(s0,b,b)=b3(bs0)3(8(q1)(q2+4q+1)44q(q1)(q2+4q+1)2(q3+3q2q+1)4(q1)q44q4(q1)q4(1q)2)\nabla^{2}f(s_{0},b,b)=\frac{b^{3}}{(b-s_{0})^{3}}\left(\begin{array}[]{ccc}8&(-q-1)\left(-q^{2}+4q+1\right)&4-4q\\ (-q-1)\left(-q^{2}+4q+1\right)&2\left(q^{3}+3q^{2}-q+1\right)&4(q-1)q\\ 4-4q&4(q-1)q&4(1-q)^{2}\\ \end{array}\right) (3)

When bc>9bc>9 we have that q<1/3q<1/3, qq is decreasing in the product of bc>9bc>9, and it remains positive. The eigenvalues of the matrix that depends only on qq and for any q(0,1/3]q\in(0,1/3] can be obtained using a closed formula (they are real roots of a degree-3 polynomial). In Figure 5 we depict their behavior. Since all eigenvalues are all positive, the candidate optimizer is indeed a minimizer.

Refer to caption
Figure 5: The eigenvalues of matrix (3) as a function of q(0,1/3]q\in(0,1/3] (and scaled by b3/(bs0)3b^{3}/(b-s_{0})^{3}).

C.5 Lemma 3.7

Proof: [Proof of Lemma 3.7] By Theorem 3.3, we know the optimizers to NLPc\textsc{NLP}_{c}^{\infty}; s=σ/c,r=ρ/c,k=κ/c,s=\sigma/c,r=\rho/c,k=\kappa/c,. These optimizers satisfy the speed bound constraints s,r,kbs,r,k\leq b as long as max{s,r,k}b\max\{s,r,k\}\leq b, i.e. ρ/cb\rho/c\leq b. Hence, when cρ/bc\geq\rho/b, Non Linear Programs NLPc\textsc{NLP}_{c}^{\infty}, NLPcb\textsc{NLP}_{c}^{b} have the same optimizers. ∎

C.6 Lemma 3.8

Proof: [Proof of Lemma 3.8] First, we observe that constraint rbr\leq b is tight for the provable optimizers for all c,bc,b when cb[9,γ1]{γ2}cb\in[9,\gamma_{1}]\cup\{\gamma_{2}\}. As the only other constraint that switches from being tight to non-tight in the same interval is kbk\leq b, we are motivated to maintain tightness for constraints rbr\leq b and the time constraint.

Given that speed ss is chosen (to be determined later), we fix r=br=b, and we set k=k2(c,b)k=k_{2}(c,b) where

k2(c,b):=2bsbcsbcs2sk_{2}(c,b):=\frac{2bs}{bcs-b-cs^{2}-s}

so as to satisfy the time constraint tightly (by solving T(s,r,b)=cT\left({s,r,b}\right)=c for kk). It remains to determine values for speed ss. To this end, we heuristically require that s=s2(c,b)s=s_{2}(c,b) where

s2(c,b):=αc+βs_{2}(c,b):=\alpha\cdot c+\beta

for some constants α,β\alpha,\beta that we allow to depend on bb. In what follows we abbreviate s2(c,b)s_{2}(c,b) by s2(c)s_{2}(c). Let s1(c),s3s_{1}(c),s_{3} be the chosen values for speed ss as summarized by the statement of Theorem 3.4 when cbγ1cb\leq\gamma_{1} and cbγ2cb\geq\gamma_{2}, respectively. We require that

s2(γ1/b)=s1(γ1/b),s2(γ2/b)=s3(γ2/b)s_{2}(\gamma_{1}/b)=s_{1}(\gamma_{1}/b),~~s_{2}(\gamma_{2}/b)=s_{3}(\gamma_{2}/b)

inducing a linear system on α,β\alpha,\beta. By solving the linear system, we obtain

α\displaystyle\alpha =b2(2γ1σ+γ2γ1γ1210γ1+9γ23γ2)2γ1(γ1γ2)γ2\displaystyle=\frac{b^{2}\left(-2\gamma_{1}\sigma+\gamma_{2}\gamma_{1}-\sqrt{\gamma_{1}^{2}-10\gamma_{1}+9}\gamma_{2}-3\gamma_{2}\right)}{2\gamma_{1}\left(\gamma_{1}-\gamma_{2}\right)\gamma_{2}}
β\displaystyle\beta =b(2γ12σ+γ22γ1γ1210γ1+9γ223γ22)2γ1γ2(γ2γ1)\displaystyle=\frac{b\left(-2\gamma_{1}^{2}\sigma+\gamma_{2}^{2}\gamma_{1}-\sqrt{\gamma_{1}^{2}-10\gamma_{1}+9}\gamma_{2}^{2}-3\gamma_{2}^{2}\right)}{2\gamma_{1}\gamma_{2}\left(\gamma_{2}-\gamma_{1}\right)}

Using the known values for γ1,γ2,σ\gamma_{1},\gamma_{2},\sigma, we obtain s2(c)=0.532412b0.0262661b2cs_{2}(c)=0.532412b-0.0262661b^{2}c, as promised. It remains to argue that s2(c,b)s_{2}(c,b), together with r=br=b, and k2(c,b)k_{2}(c,b) are feasible when γ1<cb<γ2\gamma_{1}<cb<\gamma_{2}.

The fact that s2(c)s_{2}(c) complies with bounds 0sb0\leq s\leq b follows immediately, since s2(c)s_{2}(c) is a linear strictly decreasing function in cc, and both s2(γ1/b),s2(γ2/b)s_{2}(\gamma_{1}/b),s_{2}(\gamma_{2}/b) satisfy the bounds by construction. We are therefore left with checking that 0k2(c,b)b0\leq k_{2}(c,b)\leq b which is equivalent to that

bcsbcs23s0\displaystyle~~bcs-b-cs^{2}-3s\geq 0
\displaystyle\Leftrightarrow b(bc(bc(0.001702680.000689908bc)+0.327748)1.59724)0\displaystyle~~b(bc(bc(0.00170268\,-0.000689908bc)+0.327748)-1.59724)\geq 0

Define degree-2 polynomial function g(x)=x(x(0.001702680.000689908x)+0.327748)1.59724g(x)=x(x(0.00170268\,-0.000689908x)+0.327748)-1.59724 and observe that it sufficies to prove that g(x)0g(x)\geq 0 for all x(γ1,γ2)x\in(\gamma_{1},\gamma_{2}). The roots of gg can be numerically computed as 22.8094,,5.0074,20.2699-22.8094,,5.0074,20.2699, proving that gg preserves positive sign in (γ1,γ2)(\gamma_{1},\gamma_{2}) as wanted.

Finally, the claims regarding the induced energy and competitive ratio is implied by Theorem 3.2 and obtained by evaluating the given choices of s,r,ks,r,k in E(s,r,k)E\left({s,r,k}\right). ∎

Appendix D Proofs Omitted from Section 4.

D.1 Proposition 4.2

Proof: [Proof of Proposition 4.2] The exploration algorithm explicitly ensures that round k1k-1 ends (and hence that round kk begins) at the claimed position, so only the time needs to be proved. Note that the statement is true for both robots when k=0k=0. Suppose that when LL starts its (k1)(k-1)-th round, it is at position 4k1-4^{k-1} at time 34k13\cdot 4^{k-1}. The round ends at time 34k1+4k1+1s4ks1s+(4k4ks1s)=34k3\cdot 4^{k-1}+4^{k-1}+\frac{1}{s}\cdot 4^{k}\cdot\frac{s}{1-s}+(4^{k}-4^{k}\cdot\frac{s}{1-s})=3\cdot 4^{k}, and the kk-th round will start at the claimed time. The proof is identical for RR. ∎

D.2 Lemma 4.3

Proof: [Proof of Lemma 4.3] Case 1: dD1d\in D_{1}. Let Y={L,R}{X}Y=\{L,R\}\setminus\{X\} be the robot other than XX. Observe that by Proposition 4.2, XX finds the exit at time 3K+K+d/s=4K+d/s3K+K+d/s=4K+d/s. Then XX goes towards YY at speed s=d4Kd/ss^{\prime}=\frac{d}{4K-d/s}, and so it will reach position 0 at time 4K+d/s+4Kd/sdd=8K4K+d/s+\frac{4K-d/s}{d}d=8K. We know that at time 6K6K, robot YY is starting a round at position 2K2K or 2K-2K, then goes towards 0 at full speed. Hence YY is at position 0 at time 8K8K, where it meets XX.

Case 2: dD2=[4Ks/(s+1),4Ks/(1s)]d\in D_{2}=[4Ks/(s+1),4Ks/(1-s)]. As before, the exit is found at time 4K+d/s4K+d/s. Assume for simplicity that X=LX=L (the case X=RX=R is identical by symmetry). After finding the exit at position d-d, LL goes full speed to the right. Thus at time t=4K+d/s+d+d+ds4Ks1st=4K+d/s+d+\frac{d+ds-4Ks}{1-s}, it arrives at position d+ds4Ks1s\frac{d+ds-4Ks}{1-s}. We show that at this time, RR is in its second phase and is at this position. Notice that

t\displaystyle t =4K+d/s+d+d+ds4Ks1s\displaystyle=4K+d/s+d+\frac{d+ds-4Ks}{1-s}
=8K+d/s+d4K+d+ds4Ks1s\displaystyle=8K+d/s+d-4K+\frac{d+ds-4Ks}{1-s}
=8K+d/s+d4K1s\displaystyle=8K+\frac{d/s+d-4K}{1-s}
8K(1+s/(1s)2)\displaystyle\leq 8K(1+s/(1-s)^{2})

the latter inequality being obtained from d4Ks/(1s)d\leq 4Ks/(1-s). Now, RR enters its second traveltravel phase when at position 0 at time 8K8K, and the phase ends a time 8K+1/s8Ks/(1s)=8K(1+1/(1s))8K+1/s\cdot 8Ks/(1-s)=8K(1+1/(1-s)). Since s1/2s\leq 1/2, we get 8K(1+1/(1s))8K(1+s/(1s)2)t8K(1+1/(1-s))\geq 8K(1+s/(1-s)^{2})\geq t. Therefore RR is still in its second phase at time tt, and it follows that its position at this time is s(t8K)=d+ds4Ks1ss(t-8K)=\frac{d+ds-4Ks}{1-s}. Hence LL and RR meet. It is straightforward to see that LL and RR could not have met before time tt, and thus LL and RR meet at the claimed time and position. Case 3: dD3=[4Ks/(1s),4K]d\in D_{3}=[4Ks/(1-s),4K]. Again assume that X=LX=L. This time LL finds the exit at position d-d at time 3K+K+1s4Ks1s+d4Ks1s=8K+d3K+K+\frac{1}{s}\cdot\frac{4Ks}{1-s}+d-\frac{4Ks}{1-s}=8K+d. Going full speed to the right, at time t=8K+2d+2ds/(1s)t=8K+2d+2ds/(1-s), it reaches position 2ds/(1s)2ds/(1-s). Since d4Kd\leq 4K, we have t8K(2+s/(1s))=8K(1+1/(1s))t\leq 8K(2+s/(1-s))=8K(1+1/(1-s)). As in the previous case, the second phase of RR ends at time 8K(1+1/(1s))t8K(1+1/(1-s))\geq t. Thus at time tt, RR is at position s(t8K)=2ds+2ds2/(1s)=2ds/(1s)s(t-8K)=2ds+2ds^{2}/(1-s)=2ds/(1-s). Again, one can check that LL and RR could not have met before, which concludes the proof. ∎

D.3 Lemma 4.4

Proof: [Proof of Lemma 4.4] Let XX be the robot that finds the exit at distance dd, and let K:=K(X,k)K:=K(X,k). We show that all speeds are at most 1, as well as that the evacuation time is at most 9d9d. There are three cases to consider.

Case 1: dD1d\in D_{1}. By Lemma 4.3, after meeting, both robots need to travel distance dd and have 9d8K9d-8K time remaining. By the last line of the exit phase algorithm, they go back at speed sb1:=d9d8Ks_{b1}:=\frac{d}{9d-8K} and make it in time, provided that speed sb1s_{b1} is achievable, i.e. 0<sb110<s_{b1}\leq 1. Clearly sb1>0s_{b1}>0 since 9d8K>09d-8K>0. If we assume d9d8K>1\frac{d}{9d-8K}>1, we get d>9d8Kd>9d-8K, leading to d<Kd<K, a contradiction.

Case 2: dD2d\in D_{2}. By Lemma 4.3, the robots meet at position pp such that |p|=d+ds4Ks1s|p|=\frac{d+ds-4Ks}{1-s} at time t=8K+d+d/s4K1st=8K+\frac{d+d/s-4K}{1-s}. The robots use the smallest speed sb2:=p+dt9ds_{b2}:=\frac{p+d}{t-9d} that allows the them to reach the exit in time 9d9d. We must check that 0<sb210<s_{b2}\leq 1. We argue that if the two robots used speed 11 to get to the exit after meeting, they would make it before time 9d9d. Since sb2s_{b2} allows the robots to reach the exit in time exactly 9d9d, it follows that 0<sb210<s_{b2}\leq 1.

First note that since d4Ks/(1s)d\leq 4Ks/(1-s), we have 4Kd(1s)/s4K\geq d(1-s)/s. Using speed 11 from the point they meet, the robots would reach the exit at time

8K+d+d/s4K1s+d+d+ds4Ks1s\displaystyle 8K+\frac{d+d/s-4K}{1-s}+d+\frac{d+ds-4Ks}{1-s} =4K(21+s1s)+d(1+2+s+1/s1s)\displaystyle=4K\left(2-\frac{1+s}{1-s}\right)+d\left(1+\frac{2+s+1/s}{1-s}\right)
=4K13s1s+d3s+1s(1s)\displaystyle=4K\cdot\frac{1-3s}{1-s}+d\cdot\frac{3s+1}{s(1-s)}
d1ss13s1s+d3s+1s(1s)\displaystyle\leq d\cdot\frac{1-s}{s}\cdot\frac{1-3s}{1-s}+d\cdot\frac{3s+1}{s(1-s)}
=d(13ss+3s+1s(1s))\displaystyle=d\left(\frac{1-3s}{s}+\frac{3s+1}{s(1-s)}\right)

where we have used the fact that 13s01-3s\leq 0 in the inequality. It is straighforward to show that d(13s1s+3s+1s(1s))9dd\left(\frac{1-3s}{1-s}+\frac{3s+1}{s(1-s)}\right)\leq 9d when 1/3s1/21/3\leq s\leq 1/2, proving our claim.

Case 3: dD3d\in D_{3}. Again according to Lemma 4.3, the robots meet at position pp satisfying |p|=2ds/(1s)|p|=2ds/(1-s) at time t=8K+2d+2ds/(1s)t=8K+2d+2ds/(1-s). The robots go towards the exit at speed sc3:=p+d9dts_{c3}:=\frac{p+d}{9d-t}. As in the previous case, we show that sc3s_{c3} is a valid speed by arguing that the robots have enough time if they used their full speed. If they do use speed 11 after they meet, they reach the exit at time t=8K+3d+4ds/(1s)t^{\prime}=8K+3d+4ds/(1-s). Since d4Ks/(1s)d\geq 4Ks/(1-s), we have Kd(1s)/(4s)K\leq d(1-s)/(4s). Therefore t8(d(1s)/(4s))+3d+4ds/(1s)=d3s2s+2s(1s)t^{\prime}\leq 8\cdot(d(1-s)/(4s))+3d+4ds/(1-s)=d\cdot\frac{3s^{2}-s+2}{s(1-s)}. One can check that this is 9d9d or less whenever 1/3s1/21/3\leq s\leq 1/2. ∎

D.4 Lemma 4.5

Proof: [Proof of Lemma 4.5] For any K:=4iK^{\prime}:=4^{i} power of 44 with i1i\geq 1, define BL(K,s)B_{L}(K^{\prime},s) as the energy spent by LL after reaching position K-K^{\prime} for the first time without having found the exit, ignoring the initial 1/91/9 energy spent to get to position 11. The quantity BL(K,s)B_{L}(K^{\prime},s) is the sum of energy spent in each of the first i1i-1 rounds, and so

BL(K,s)\displaystyle B_{L}(K^{\prime},s) =j=0i1(4j+s2(4j+1s/(1s))+4j+14j+1s/(1s))\displaystyle=\sum_{j=0}^{i-1}\left(4^{j}+s^{2}(4^{j+1}s/(1-s))+4^{j+1}-4^{j+1}s/(1-s)\right)
=j=0i1(4j+1(5/4+(s3s)/(1s)))\displaystyle=\sum_{j=0}^{i-1}\left(4^{j+1}(5/4+(s^{3}-s)/(1-s))\right)
=4/3(4i1)(5/4s(s+1))\displaystyle=4/3(4^{i}-1)(5/4-s(s+1))
=1/3(K1)(54s(s+1))\displaystyle=1/3(K^{\prime}-1)(5-4s(s+1))

We define BR(2K,s)B_{R}(2K^{\prime},s) similarly for RR, i.e. BR(2K,s)B_{R}(2K^{\prime},s) is the energy spent by RR when its (i1)(i-1)-th round is finished and it reached position 2K2K^{\prime} for the first time, ignoring the initial 2/92/9 energy to get at position 22. We get

BR(2K,s)\displaystyle B_{R}(2K^{\prime},s) =j=0i1(24j+s2(24j+1s/(1s))+24j+124j+1s/(1s))\displaystyle=\sum_{j=0}^{i-1}\left(2\cdot 4^{j}+s^{2}(2\cdot 4^{j+1}s/(1-s))+2\cdot 4^{j+1}-2\cdot 4^{j+1}s/(1-s)\right)
=2BL(K,s)=2/3(K1)(54s(s+1))\displaystyle=2B_{L}(K^{\prime},s)=2/3(K^{\prime}-1)(5-4s(s+1))

We may now calculate the three possible cases of energy. Assume that X{L,R}X\in\{L,R\} finds the exit and Y={L,R}{X}Y=\{L,R\}\setminus\{X\}. Observe that

BX(K,s)+BY(2K,s)=(K1)(54s(s+1))=F(K,s)B_{X}(K,s)+B_{Y}(2K,s)=(K-1)(5-4s(s+1))=F(K,s)

We implictly use Lemma 4.3 for the distance traveled by XX to catch up to YY after finding the exit, and the distance traveled back by both robots. In the E(K,d,s)E(K,d,s) expressions that follow, for clarity we partition the terms into 3 brackets, which respectively represent the energy spent by XX to find the exit and catch up to YY, the energy spent by YY before being caught, and the energy spent by both robots to go to the exit.

Case 1: dD1d\in D_{1}. The total energy spent is

E(K,d,s)\displaystyle E(K,d,s) =[BX(K,s)+K+s2d+sc12d]+[BY(2K,s)+2K]+[2sb12d]\displaystyle=\left[B_{X}(K,s)+K+s^{2}d+s_{c1}^{2}d\right]+\left[B_{Y}(2K,s)+2K\right]+\left[2s_{b1}^{2}d\right]
=F(K,s)+3K+d(s2+sc12+2sb12)\displaystyle=F(K,s)+3K+d(s^{2}+s_{c1}^{2}+2s_{b1}^{2})

Case 2: dD2d\in D_{2}. In this case, the energy spent is

E(K,d,s)\displaystyle E(K,d,s) =[BX(K,s)+K+s2d+d+d+ds4Ks1s]+[BY(2K,s)+2K+s2(d+ds4Ks1s)]+\displaystyle=\left[B_{X}(K,s)+K+s^{2}d+d+\frac{d+ds-4Ks}{1-s}\right]+\left[B_{Y}(2K,s)+2K+s^{2}\left(\frac{d+ds-4Ks}{1-s}\right)\right]+
[2sb22(d+d+ds4Ks1s)]\displaystyle\quad\left[2s_{b2}^{2}\left(d+\frac{d+ds-4Ks}{1-s}\right)\right]
=F(K,s)+3K+(2d4Ks1s)(1+s2+2sb22)\displaystyle=F(K,s)+3K+\left(\frac{2d-4Ks}{1-s}\right)(1+s^{2}+2s_{b2}^{2})

Case 3: dD3d\in D_{3}. The energy spent is

E(K,d,s)\displaystyle E(K,d,s) =[BX(K,s)+K+s24Ks/(1s)+d4Ks/(1s)+d+2ds/(1s)]+\displaystyle=\left[B_{X}(K,s)+K+s^{2}4Ks/(1-s)+d-4Ks/(1-s)+d+2ds/(1-s)\right]+
[BY(2K,s)+2K+s2(2ds/(1s))]+[2sb32(d+2ds/(1s))]\displaystyle\quad\left[B_{Y}(2K,s)+2K+s^{2}(2ds/(1-s))\right]+\left[2s^{2}_{b3}\left(d+2ds/(1-s)\right)\right]
=F(K,s)+3K+4Ks/(1s)(s21)+d(s1s(2+2s2+2sb32)+2+2sb32)\displaystyle=F(K,s)+3K+4Ks/(1-s)(s^{2}-1)+d\left(\frac{s}{1-s}(2+2s^{2}+2s_{b3}^{2})+2+2s_{b3}^{2}\right)
=F(K,s)+3K4Ks(s+1)+2d1s(s3+sb32(s+1)+1)\displaystyle=F(K,s)+3K-4Ks(s+1)+\frac{2d}{1-s}(s^{3}+s_{b3}^{2}(s+1)+1)

D.5 Lemma 4.6

Proof: [Proof of Lemma 4.6] We analyze each case separately.

Case 1: dD1=[K,4Ks/(s+1)]d\in D_{1}=[K,4Ks/(s+1)]. According to Lemma 4.5, we have

E(K,d,s)/d\displaystyle E(K,d,s)/d =1/d((K1)(54s(s+1))+3K+d(s2+sc12+2sb12))\displaystyle=1/d\cdot((K-1)(5-4s(s+1))+3K+d(s^{2}+s_{c1}^{2}+2s_{b1}^{2}))
=(K1)(54s(s+1))+3Kd+s2+(d4Kd/s)2+2(d9d8K)2\displaystyle=\dfrac{(K-1)(5-4s(s+1))+3K}{d}+s^{2}+\left(\frac{d}{4K-d/s}\right)^{2}+2\left(\frac{d}{9d-8K}\right)^{2}

Consider the case d=Kd=K. Plugging in s=0.39403s=0.39403, the above evaluates to 8.42587140912.8028414364/K8.4258714091-2.8028414364/K. We claim that E(K,d,s)/dE(K,d,s)/d is a decreasing function over interval D1D_{1}, and therefore attains its maximum when d=Kd=K. Assuming this is true, adding the initialization energy of 1/31/3 omitted so far and given that dKd\geq K, the energy ratio is at most

8.42587140912.8028414364/K+1/(3K)8.425888.4258714091-2.8028414364/K+1/(3K)\leq 8.42588

We now prove that E(K,d,s)/dE(K,d,s)/d is decreasing over the interval D1D_{1}. Let

f1(K,d,s):=\displaystyle f_{1}(K,d,s):= d2(9d8K)2\displaystyle\frac{d^{2}}{(9d-8K)^{2}}
f2(K,d,s):=\displaystyle f_{2}(K,d,s):= d2(d4Ks)2\displaystyle\frac{d^{2}}{(d-4Ks)^{2}}
f3(K,d,s):=\displaystyle f_{3}(K,d,s):= 4K(s2+s2)+4s(s+1)5d,\displaystyle\frac{-4K\left(s^{2}+s-2\right)+4s(s+1)-5}{d},

and observe that

E(K,d,s)d=2f1(K,d,s)+s2f2(K,d,s)+f3(K,d,s)+s2.\frac{E(K,d,s)}{d}=2f_{1}(K,d,s)+s^{2}f_{2}(K,d,s)+f_{3}(K,d,s)+s^{2}.

The plan is to prove that

dE(K,d,s)/d<0.\frac{\partial}{\partial d}E(K,d,s)/d<0.

For this we calculate

df1(K,d,s):=\displaystyle\frac{\partial}{\partial d}f_{1}(K,d,s):= 16dK(9d8K)3\displaystyle-\frac{16dK}{(9d-8K)^{3}}
df2(K,d,s):=\displaystyle\frac{\partial}{\partial d}f_{2}(K,d,s):= 8dKs(d4Ks)3\displaystyle-\frac{8dKs}{(d-4Ks)^{3}}
df3(K,d,s):=\displaystyle\frac{\partial}{\partial d}f_{3}(K,d,s):= 4K(s2+s2)4s(s+1)+5d2,\displaystyle\frac{4K\left(s^{2}+s-2\right)-4s(s+1)+5}{d^{2}},

Now we claim that all dfi(K,d,s)\frac{\partial}{\partial d}f_{i}(K,d,s) are increasing functions in dd, for i=1,2,3i=1,2,3. Indeed, first,

2d2f1(K,d,s)=32K(9d+4K)(9d8K)4>0\frac{\partial^{2}}{\partial d^{2}}f_{1}(K,d,s)=\frac{32K(9d+4K)}{(9d-8K)^{4}}>0

since dKd\geq K. Hence df1(K,d,s)\frac{\partial}{\partial d}f_{1}(K,d,s) is increasing in dd.

Second,

2d2f2(K,d,s)=16Ks(d+2Ks)(d4Ks)4\frac{\partial^{2}}{\partial d^{2}}f_{2}(K,d,s)=\frac{16Ks(d+2Ks)}{(d-4Ks)^{4}}

is positive (and well defined), since d4ks/(1+s)d\leq 4ks/(1+s). Hence df2(K,d,s)\frac{\partial}{\partial d}f_{2}(K,d,s) is increasing in dd.

Third, we show that df3(K,d,s)\frac{\partial}{\partial d}f_{3}(K,d,s) is increasing in dd. For this it is enough to prove that 4K(s2+s2)4s(s+1)+5<0.4K\left(s^{2}+s-2\right)-4s(s+1)+5<0. For s=0.39403s=0.39403 (and in fact for all s(2,1)s\in(-2,1)) the strict inequality can be written as K4s2+4s54s2+4s8,K\geq\frac{4s^{2}+4s-5}{4s^{2}+4s-8}, which we show next it is satisfied. Indeed, it is easy to see that 4s2+4s54s2+4s85/8\frac{4s^{2}+4s-5}{4s^{2}+4s-8}\leq 5/8 (which is attained for s=0s=0), while K4K\geq 4, hence the claim follows.

To resume, we showed that dfi(K,d,s)\frac{\partial}{\partial d}f_{i}(K,d,s) are increasing functions in dd, for i=1,2,3i=1,2,3. Recalling that s=0.39403s=0.39403, and since d4ks/(1+s)d\leq 4ks/(1+s), we obtain that

dE(K,d,s)/d\displaystyle\frac{\partial}{\partial d}E(K,d,s)/d
2df1(K,4Ks/(1+s),s)+s2df2(K,4Ks/(1+s),s)+df3(K,4Ks/(1+s),s)\displaystyle\leq 2\frac{\partial}{\partial d}f_{1}(K,4Ks/(1+s),s)+s^{2}\frac{\partial}{\partial d}f_{2}(K,4Ks/(1+s),s)+\frac{\partial}{\partial d}f_{3}(K,4Ks/(1+s),s)
=(s+1)2(s(4K(s(s+1)(49s(7s6)+76)8)s(7s(28s(7s+1)365)+1774)+452)40)16K2s2(7s2)3\displaystyle=\frac{(s+1)^{2}(s(4K(s(s+1)(49s(7s-6)+76)-8)-s(7s(28s(7s+1)-365)+1774)+452)-40)}{16K^{2}s^{2}(7s-2)^{3}}
=2.192621.79465KK2.\displaystyle=\frac{2.19262\,-1.79465K}{K^{2}}.

Since K4K\geq 4, the latter quantity is clearly negative. This shows that dE(K,d,s)/d\frac{\partial}{\partial d}E(K,d,s)/d is negative (in the given domain), hence E(K,d,s)/dE(K,d,s)/d is decreasing in dd.


Case 2: dD2=[4Ks/(s+1),4Ks/(1s)]d\in D_{2}=[4Ks/(s+1),4Ks/(1-s)]. In this case, the energy ratio E(K,d,s)/dE(K,d,s)/d is

1/d(1/3+(K1)(54s(s+1))+3K+(2d4Ks1s)(1+s2+2sb22))\displaystyle\quad 1/d\cdot\left(1/3+(K-1)(5-4s(s+1))+3K+\left(\frac{2d-4Ks}{1-s}\right)(1+s^{2}+2s_{b2}^{2})\right)
=(K1)(54s(s+1))+3Kd+(2d4Ksd(1s))(1+s2+2(2d4Ksd(89s1/s)+4K(2s1))2)\displaystyle=\frac{(K-1)(5-4s(s+1))+3K}{d}+\left(\frac{2d-4Ks}{d(1-s)}\right)\left(1+s^{2}+2\left(\frac{2d-4Ks}{d(8-9s-1/s)+4K(2s-1)}\right)^{2}\right)

We will show that this expression achieves its maximum at d=4Ks/(1s)d=4Ks/(1-s). When s=0.39403s=0.39403, then above yields 8.4257860601.0776069241/K8.425786060-1.0776069241/K. Given that 1/(3d)1/(3.39186K)1/(3d)\leq 1/(3.39186K), this implies that the energy ratio is at most

8.4257860601.0776069241/K+1/(3.39186K)8.425888.425786060-1.0776069241/K+1/(3.39186K)\leq 8.42588

We prove that E(K,d,s)/dE(K,d,s)/d is an increasing function over interval D2D_{2}. First we compute dE(K,d,s)d\frac{\partial}{\partial d}\frac{E(K,d,s)}{d} and we substitute s=0.39403s=0.39403 to find 1d2(d0.442471K)3g(K,d)\frac{1}{d^{2}(d-0.442471K)^{3}}g(K,d), where

g(K,d):=\displaystyle g(K,d):= d3(7.8438K+2.80279)+d2K(15.5674K3.72046)\displaystyle d^{3}(7.8438K+2.80279)+d^{2}K(-15.5674K-3.72046)
+dK2(8.91957K+1.6462)+K3(1.31555K0.242798).\displaystyle+dK^{2}(8.91957K+1.6462)+K^{3}(-1.31555K-0.242798).

Note that d4Ks/(1+s)1.13064Kd\geq 4Ks/(1+s)\approx 1.13064K, and hence d2(d0.442471K)3>0d^{2}(d-0.442471K)^{3}>0 for all values of dd under consideration. Therefore the lemma will follow if we show that g(K,d)0g(K,d)\geq 0 as well.

g(K,d)g(K,d) is a degree-3 polynomial with positive leading coefficient. It attains a local minimum at the largest real root of

dg(K,d)=3d2(7.8438K+2.80279)+2dK(15.5674K3.72046)+K2(8.91957K+1.6462)\frac{\partial}{\partial d}g(K,d)=3d^{2}(7.8438K+2.80279)+2dK(-15.5674K-3.72046)+K^{2}(8.91957K+1.6462)

which is

d0(K):=K(0.661559K+0.0212482K(129.818K+8.39821)+0.158106)K+0.357325d_{0}(K):=\frac{K\left(0.661559K+0.0212482\sqrt{K(129.818K+8.39821)}+0.158106\right)}{K+0.357325}

Now we observe that for all K>0K>0, we have d0(K)<4Ks/(1+s)1.13064Kd_{0}(K)<4Ks/(1+s)\approx 1.13064K.

From the above, it follows that g(K,d)g(K,d) is monotonically increasing for d4Ks/(1+s)d\geq 4Ks/(1+s), and therefore

g(K,d)g(K,4Ks/(1+s))=(0.205742K+0.913416)K30g(K,d)\geq g(K,4Ks/(1+s))=(0.205742K+0.913416)K^{3}\geq 0

as wanted.


Case 3: dD3=[4Ks/(1s),4K]d\in D_{3}=[4Ks/(1-s),4K]. The energy ratio E(K,d,s)/dE(K,d,s)/d is

1/d(1/3+(K1)(54s(s+1))+3K4Ks(s+1)+2d1s(s3+sb32(s+1)+1))\displaystyle\quad 1/d\cdot\left(1/3+(K-1)(5-4s(s+1))+3K-4Ks(s+1)+\frac{2d}{1-s}(s^{3}+s_{b3}^{2}(s+1)+1)\right)
=1/3+(K1)(54s(s+1))+3K4Ks(s+1)d+2s3+21s+2s+21s(d(1+s)d(79s)+8K(s1))2\displaystyle=\frac{1/3+(K-1)(5-4s(s+1))+3K-4Ks(s+1)}{d}+\frac{2s^{3}+2}{1-s}+\frac{2s+2}{1-s}\left(\frac{d(1+s)}{d(7-9s)+8K(s-1)}\right)^{2}

In this case, we claim that this expression is decreasing over D3D_{3} and achieves its maximum at d=4Ks/(1s)d=4Ks/(1-s). When s=0.39403s=0.39403, the above gives 8.4257860601.0776069241/K8.425786060-1.0776069241/K (which is the same as in case 2, as one should expect). Given that 1/(3d)1/(7.80296K)1/(3d)\leq 1/(7.80296K), we get that the energy ratio is at most

8.4257860601.0776069241/K+1/(7.80296K)8.425888.425786060-1.0776069241/K+1/(7.80296K)\leq 8.42588

Let us prove that E(K,d,s)/dE(K,d,s)/d is indeed decreasing. Note that E(K,d,s)/dE(K,d,s)/d equals

2(s+1)31sg(K,d,s)+h(K,d,s)+2(s3+1)1s\frac{2(s+1)^{3}}{1-s}g(K,d,s)+h(K,d,s)+\frac{2\left(s^{3}+1\right)}{1-s}

where

g(K,d,s):=d2(d(79s)+8K(s1))2,h(K,d,s):=8K(s2s+1)+4s(s+1)5d.g(K,d,s):=\frac{d^{2}}{(d(7-9s)+8K(s-1))^{2}},~~h(K,d,s):=\frac{8K\left(-s^{2}-s+1\right)+4s(s+1)-5}{d}.

In what follows we prove that both g(K,d,s),h(K,d,s)g(K,d,s),h(K,d,s) are strictly decreasing when d4Ks/(1s)d\geq 4Ks/(1-s), implying the claim of the lemma.

First we show that h(K,d,s)h(K,d,s) is decreasing. For that note that, using the fixed value of s=0.39403s=0.39403, we have h(K,d,s)=(3.60558K2.80279)/dh(K,d,s)=(3.60558K-2.80279)/d, and the latter expression (in dd) is clearly strictly decreasing for all constants K>1K>1.

Now we show that g(K,d,s)g(K,d,s) is strictly decreasing for all d4Ks/(1s)d\geq 4Ks/(1-s). For that observe that for the specific constant ss, and since d4Ks/(1s)2.60107Kd\geq 4Ks/(1-s)\approx 2.60107K, we have that

|d(79s)+8K(s1)|=d(79s)+8K(s1)>0.|d(7-9s)+8K(s-1)|=d(7-9s)+8K(s-1)>0.

Hence, to show that g(K,d,s)g(K,d,s) is strictly decreasing, it is enough to prove that

q(K,d,s):=d(d(79s)+8K(s1))q(K,d,s):=\frac{d}{(d(7-9s)+8K(s-1))}

is strictly decreasing in d4Ks/(1s)d\geq 4Ks/(1-s). First observe that the rational function is well defined for these values of dd, since the denominator becomes 0 only when d=8K(s1)9s7<4Ks/(1s)d=\frac{8K(s-1)}{9s-7}<4Ks/(1-s) (the last inequality is easy to verify). To that end, we compute

dq(K,d,s)=8K(s1)(d(79s)+8K(s1))2\frac{\partial}{\partial d}q(K,d,s)=\frac{8K(s-1)}{(d(7-9s)+8K(s-1))^{2}}

which is of course negative for the given value of s<1s<1.