This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Lower Bound on Observability for Target Tracking with Range Sensors and its Application to Sensor Assignment

Lifeng Zhou and Pratap Tokekar The authors are with the Department of Electrical and Computer Engineering, Virginia Tech USA. {lfzhou, tokekar}@vt.edu.This material is based upon work supported by the National Science Foundation under Grant Nos. 1566247 and 1637915.
Abstract

We study two sensor assignment problems for multi-target tracking with the goal of improving the observability of the underlying estimator. In the restricted version of the problem, we focus on assigning unique pairs of sensors to each target. We present a 1/31/3–approximation algorithm for this problem. We use the inverse of the condition number as the value function. If the target’s motion model is not known, the inverse cannot be computed exactly. Instead, we present a lower bound for range-only sensing.

In the general version, the sensors must form teams to track individual targets. We do not force any specific constraints on the size of each team, instead assume that the value function is monotonically increasing and is submodular. A greedy algorithm that yields a 1/21/2–approximation. However, we show that the inverse of the condition number is neither monotone nor submodular. Instead, we present other measures that are monotone and submodular. In addition to theoretical results, we evaluate our results empirically through simulations.

I Introduction

State estimation is a fundamental problem in sensorics and finds many applications such as localization, mapping, and target tracking [1, 2]. The estimator performance can be improved by exploiting the observability of the underlying system [3, 4, 5, 6]. We study selecting sensors to improve the observability in tracking a potentially mobile target.

Observability is a basic concept in control theory and has been widely applied in robotics. Observability for range-only beacon sensors, in particular, has been closely studied for underwater navigation. Gadre and Stilwell [3] analyzed the local and global observability [7] for the localization of an Autonomous Underwater Vehicle by an acoustic beacon. The problems of single vehicle localization and multi-vehicle relative localization are studied in [5] using an observability criterion introduced in [8]. In these works, it is the sensors that are moving. Consequently, the sensors know their control vectors and can thus, compute the observability matrix and its measures. In target tracking with fixed or mobile sensors, however, the control inputs for the targets are unknown. In recent work, Williams and Sukhatme [6] studied a multi-sensor localization and target tracking problem where they showed how to leverage graph rigidity to improve the observability for sensor team localization and robust target tracking.

While others have studied similar problems in the past [6, 5], we focus uniquely on the case when the control inputs for the target are not known to the sensors. Consequently, we cannot compute the observability matrix of the resulting system. Our first contribution is to present a novel lower bound on the observability for the case of unknown target motion tracked by range-only sensors. Specifically, we show how to lower bound the condition number [8] of the partially known observability matrix using only the known part (Section III).

We then study two sensor assignment problems. In the first problem, the goal is to assign unique pairs of sensors to a target. Our second contribution is a greedy assignment algorithm for this problem. We prove that the greedy algorithm achieves a 1/31/3–approximation of the optimal solution. We show that the greedy algorithm performs much better than 1/31/3 in practice (Section V-C). Since the optimal solution cannot be computed, in order to compare the greedy approach we also consider a relaxed version of the assignment problem which gives an upper bound for the optimal.

We then study a general assignment problem where a set of sensors are to be assigned to a specific target. There are no restrictions on the number of targets assigned to a specific sensor. Instead, we let the algorithm decide the optimal configuration of sensor teams assigned to each target. If the weight function is submodular and monotone111We use “monotone” and “monotone increasing” interchangeably., a greedy algorithm gives a 1/21/2–approximation [9]. However, our third contribution is to prove that the lower bound of the inverse of the condition number is neither submodular nor monotone. Instead, we use other observability measures such as the trace, log determinant and rank of symmetric observability matrix, and the trace of inverse symmetric observability matrix and show them to be submodular and monotone. We evaluate this algorithm through simulations where we find that sensors are assigned to targets almost uniformly (Section V-D).

II Problem Formulation for sensor assignment

We consider a scenario where there are NN sensors and LL targets in the environment. Our goal is to assign sensors to track the target. We use the notation σ(l)\sigma(l) to represent the set of sensors assigned to target tlt_{l}. Similarly, let σ1(i)\sigma^{-1}(i) give the set of targets assigned to sensor sis_{i}. We also use σi(l)\sigma_{i}(l) to give the ithi^{th} sensor assigned to tlt_{l}. We order the assigned sensors by using their IDs such that σ1(l)<σ2(l)<σ3(l)<\sigma_{1}(l)<\sigma_{2}(l)<\sigma_{3}(l)<\ldots. Let ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}), and ω(Sl,tl)\omega(S_{l},t_{l}) be some measure of the observability of tracking tlt_{l} with sis_{i} and sjs_{j}, and with a set of sensor(s) SlS_{l}, respectively.

We study the following sensor assignment problems. We start with the problem of assigning pairs of sensors to each target.222Theorems 2 and 3 show that at least two sensors are necessary when ω()\omega(\cdot) is the inverse condition number of the observability matrix.

Problem 1 (Unique Pair Assignment).

Given a set of sensor positions, 𝒮:={s0,,sN}\mathcal{S}:=\{s_{0},\ldots,s_{N}\} and a set of target estimates at time tt, 𝒯:={t0,,tL}\mathcal{T}:=\{t_{0},\ldots,t_{L}\}, find an assignment of unique pairs of sensors to targets:

maximizei=1Lω(σ1(l),σ2(l),tl)\text{maximize}\sum_{i=1}^{L}\omega(\sigma_{1}(l),\sigma_{2}(l),t_{l}) (1)

with the added constraint each sensor is assigned to at most one target. That is, for all i=1,,Ni=1,\ldots,N we have |σ1(i)|1|\sigma^{-1}(i)|\leq 1, assuming N2LN\geq 2L.

We then study the general version of the problem where each target can be tracked by more than two sensors. That is, the sensors form teams of varying sizes to track individual targets. Sensors within a team can share measurements so as to better track the targets. We constrain each sensor to be assigned to only one target, that is, communicate with only one team. This is motivated by scenarios where sensing multiple targets can be time consuming (as is the case with radio sensors [10]) or communicating multiple measurements can be time and energy consuming.

Problem 2 (General Assignment).

Given a set of sensor positions, 𝒮:={s0,,sN}\mathcal{S}:=\{s_{0},\ldots,s_{N}\} and a set of target estimates at time tt, 𝒯:={t0,,tL}\mathcal{T}:=\{t_{0},\ldots,t_{L}\}, find an assignment of sets of sensors to targets:

maximizei=1Lω(σ(l),tl)\text{maximize}\sum_{i=1}^{L}\omega(\sigma(l),t_{l}) (2)

with the added constraint each sensor is assigned to at most one target.

We present a 1/31/3–approximation to solve Problem 1 for any ω()\omega(\cdot). Problem 2 is the more general assignment problem which is difficult to solve for arbitrary ω()\omega(\cdot). However, for the specific class of submodular functions, there exists a 1/21/2–approximation by a greedy algorithm [9]. Submodularity captures the notion of diminishing returns, i.e., the marginal gain of assigning an additional sensor diminishes as more and more sensors are assigned to track the same target.

A typical measure of observability is the condition number of the observability matrix. When the target is moving, the condition number of the observability matrix cannot be computed, since the control input for the target is unknown to the sensors. We find a lower bound on the inverse of the condition number of the observability matrix. We use this lower bound as our measure, i.e., ω()\omega(\cdot), to find the assignments in Problem 1. However, we show that this measure is not submodular (in fact, not even monotone increasing). Instead, we can use a number of other measures (e.g., the trace of symmetric observability matrix, the trace of inverse symmetric observability matrix, log determinant and rank of the symmetric observability matrix) which we know to be submodular and monotone increasing (Theorem 8). We start by how to bound the inverse of condition number.

III Bounding the Observability

Consider a mobile target whose position is denoted by oo. Suppose there are NN stationary sensors that can measure the distance333We use the square of the distance/range for mathematical convenience. to the target. We have:

{o˙=uo,zi=hi(o)=12pio22,i=1,,N\left\{\begin{array}[]{ll}\dot{o}=u_{o},\\ z_{i}=h_{i}(o)=\frac{1}{2}\|p_{i}-o\|_{2}^{2},~i=1,...,N\end{array}\right. (3)

where o:=[ox,oy]To:=[o_{x},o_{y}]^{T} gives the 2D position of the target, and uo:=[uox,uoy]Tu_{o}:=[u_{ox},u_{oy}]^{T} defines its control input, which is unknown. We assume an upper bound on the control input, given by uo,max=maxuo2u_{o,\max}=\max\|u_{o}\|_{2}. ziz_{i} defines the range-only measurement from each sensor sis_{i} whose position is given by pi=[pix,piy]Tp_{i}=[p_{ix},p_{iy}]^{T}. For simplicity, we also assume that the target does not collide with any sensor, i.e., pio20\|p_{i}-o\|_{2}\neq 0 and no two sensors are deployed at the same position.

We analyze the weak local observability matrix, O(o,uo)O(o,u_{o}), of this multi-sensor target tracking system. We show how to lower bound the inverse of the condition number of O(o,uo)O(o,u_{o}), given by C1(O(o,uo))C^{-1}(O(o,u_{o})), independent of uou_{o}. We also show that the lower bound, C¯1(Oi(o,uo))\underline{C}^{-1}(O_{i}(o,u_{o})), is tight.

We compute the local nonlinear observability matrix [7, 6] for this system (Equation 3) as,

O(o,uo)=[L0h1L1h1L0h2L1h2L0hNL1hN]=[oxp1x,oyp1yuox,uoy0,0oxp2x,oyp2yuox,uoy0,0oxpNx,oypNyuox,uoy0,0].O(o,u_{o})=\begin{bmatrix}\nabla L_{0}^{h_{1}}\\ \nabla L_{1}^{h_{1}}\\ \vdots\\ \nabla L_{0}^{h_{2}}\\ \nabla L_{1}^{h_{2}}\\ \vdots\\ \\ \vdots\\ \\ \nabla L_{0}^{h_{N}}\\ \nabla L_{1}^{h_{N}}\\ \vdots\par\end{bmatrix}=\begin{bmatrix}o_{x}-p_{1x},o_{y}-p_{1y}\\ u_{ox},u_{oy}\\ 0,0\\ \vdots\\ o_{x}-p_{2x},o_{y}-p_{2y}\\ u_{ox},u_{oy}\\ 0,0\\ \vdots\\ \\ \vdots\\ \\ o_{x}-p_{Nx},o_{y}-p_{Ny}\\ u_{ox},u_{oy}\\ 0,0\\ \vdots\par\end{bmatrix}. (4)

This equation can be rewritten as,

O(o,uo)\displaystyle O(o,u_{o}) =\displaystyle= [oxp1x,oyp1yoxp2x,oyp2yoxpNx,oypNyuox,uoy](N+1)×2.\displaystyle\begin{bmatrix}o_{x}-p_{1x},o_{y}-p_{1y}\\ o_{x}-p_{2x},o_{y}-p_{2y}\\ \vdots\\ o_{x}-p_{Nx},o_{y}-p_{Ny}\\ u_{ox},u_{oy}\end{bmatrix}_{(N+1)\times 2}. (5)

The state of the target oo is weakly locally observable if the local nonlinear observability matrix has full column rank [7]. However, the rank test for the observability of the system is a binary condition which does not tell the degree of the observability or how good the observability is. The condition number [8], defined as the ratio of the largest singular value to the smallest, can be used to measure this degree of unobservability. A larger condition number suggests worse observability. We use the inverse of condition number given as,

C1(O(o,uo))=σmin(O(o,uo))σmax(O(o,uo)).C^{-1}(O(o,u_{o}))=\frac{\sigma_{\min}(O(o,u_{o}))}{\sigma_{\max}(O(o,u_{o}))}. (6)

Note that, C1[0,1]C^{-1}\in[0,1]. C1=0C^{-1}=0 means O(o,uo)O(o,u_{o}) is singular and C1=1C^{-1}=1 means O(o,uo)O(o,u_{o}) is well conditioned. A larger C1C^{-1} means better observability (see more details in [5]).

In the local nonlinear observability matrix O(o,uo)O(o,u_{o}), uou_{o} is unknown and not controllable by the sensor. On the other hand, opio-p_{i}, depends on the relative state between each sensor sis_{i} and target oo and is known to the sensor (assuming an estimate of the target’s position is known). The system can control opio-p_{i} either by moving the sensors or assigning new sensors to track the target.

Theorem 1.

For the multi-sensor-target system (Equation 3) with the number of sensors, N2N\geq 2, the inverse of the condition number is lower bounded by σmin(O(o))σmax2(O(o))+uo,max2\dfrac{\sigma_{\min}(O(o))}{\sqrt{\sigma^{2}_{\max}(O(o))+u_{o,\max}^{2}}}.

We present the full proof for this and all other results in the in the appendix.

We wish to improve the worst case, i.e., the lower bound of C1(O(o,uo))C^{-1}(O(o,u_{o})), by optimizing the sensor-target relative state, which can be controlled by the sensor. For example, if the sensors are mobile, they can move so as to improve the lower bound. If only a subset of sensors are active at a time, we can choose the appropriate subset to improve the lower bound. In the following, we will show that at least two sensors are required to improve the lower bound.

Theorem 2.

The lower bound of the observability metric in one-sensor-target system, C¯1(Oi(o,uo))\underline{C}^{-1}(O_{i}(o,u_{o})), cannot be controlled by the sensor.

When the number of sensors, N2N\geq 2, we have a positive result that shows that the sensors can improve the lower bound on the condition number of optimizing their positions.

Theorem 3.

Suppose that the number of sensors, N2N\geq 2. Even though the contribution to the observability matrix from the target’s input, O(uo)O(u_{o}), is unknown and cannot be controlled, if the sensors increase C1(O(o))C^{-1}(O(o)) and σmin(O(o))\sigma_{\min}(O(o)) (the inverse of condition number and the smallest singular number of the relative state contribution O(o)O(o)), then the lower bound of C1(O(o,uo))C^{-1}(O(o,u_{o})) also increases.

Remark 1.

The lower bound C¯1(O(o,uo))\underline{C}^{-1}(O(o,u_{o})) is tight when the target is known to be stationary. If uo{0}u_{o}\in\{0\}, O(o,uo)=O(o)O(o,u_{o})=O(o) by Equation 5 and Theorem 1. Thus, the lower bound C¯1(O(o,uo))=C1(O(o,uo))\underline{C}^{-1}(O(o,u_{o}))={C}^{-1}(O(o,u_{o})) implying that the lower bound is tight.

Consider the special case when N=2N=2. We use ii and jj to denote the two sensors tasked with tracking the target. The local observability matrix Oi,j(o,uo)O_{i,j}(o,u_{o}) for the ijoi-j-o system is given by,

Oi,j(o,uo)=[oxpixoypiyoxpjxoypjyuoxuoy].O_{i,j}(o,u_{o})=\begin{bmatrix}o_{x}-p_{ix}&o_{y}-p_{iy}\\ o_{x}-p_{jx}&o_{y}-p_{jy}\\ u_{ox}&u_{oy}\end{bmatrix}. (7)

For ease of notation, we represent the relative sensor-target position and sensor-target orientation in polar coordinates with the target at the center:

{pixox=diocosθi,piyoy=diosinθi,pjxox=djocosθj,pjyoy=djosinθj.\left\{\begin{array}[]{ll}p_{ix}-o_{x}=d_{io}\cos\theta_{i},\\ p_{iy}-o_{y}=d_{io}\sin\theta_{i},\\ p_{jx}-o_{x}=d_{jo}\cos\theta_{j},\\ p_{jy}-o_{y}=d_{jo}\sin\theta_{j}.\end{array}\right. (8)

where diod_{io}, djod_{jo}, θi\theta_{i} and θj\theta_{j} indicate the relative distances and orientations between sensors i,ji,j and target oo as shown in Figure 1.

Refer to caption
Figure 1: Relative sensor-target position and sensor-target orientation in polar coordinates.
Theorem 4.

The lower bound of the inverse of condition number of the observability matrix for ijoi-j-o system is given by,

C¯1(Oi,j(o,uo))=\displaystyle\underline{C}^{-1}(O_{i,j}(o,u_{o}))=
λmin(OT(o)O(o))λmax(OT(o)O(o))+uo2=\displaystyle\sqrt{\frac{\lambda_{\min}(O^{T}(o)O(o))}{\lambda_{\max}(O^{T}(o)O(o))+u_{o}^{2}}}=
1+α21+α4+2α2cos(2θji)1+α2+1+α4+2α2cos(2θji)+2uo2/dio2,\displaystyle\sqrt{\frac{1+\alpha^{2}-\sqrt{1+\alpha^{4}+2\alpha^{2}\cos(2\theta_{ji})}}{1+\alpha^{2}+\sqrt{1+\alpha^{4}+2\alpha^{2}\cos(2\theta_{ji})}+2u_{o}^{2}/d_{io}^{2}}},~ (9)

where α:=djo/dio\alpha:=d_{jo}/d_{io}.

We plot C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) as a function of α[0,5]\alpha\in[0,5] and angle θji[π,π]\theta_{ji}\in[-\pi,\pi] with uo,maxu_{o,\max} and diod_{io} selected as 1m/s1~m/s and 1m1~m, respectively (Figure 2). Note that C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) reaches its maximum when α=1\alpha=1 and θji=±π2\theta_{ji}=\pm\frac{\pi}{2}. That is, both sensors are at the same distance from the target and are perpendicular with respect to the target. On the other hand, when θji=0\theta_{ji}=0, θji=π\theta_{ji}=\pi, α=0\alpha=0 or α\alpha\to\infty, C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) reaches zero. We summarize the results in the following theorem.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 2: The lower bound of the inverse of condition number of the observability matrix for the ijoi-j-o system as a function of the distance ratio, α\alpha, and angle θ\theta in Cartesian (a, b) and Polar (c, d) coordinates.
Theorem 5.

The lower bound of the inverse of the condition number for the observability matrix for ijoi-j-o system, C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})), reaches its maximum, 11+uo2/dio2\sqrt{\frac{1}{1+u_{o}^{2}/d_{io}^{2}}}, at α=1\alpha=1 and θji=±π2\theta_{ji}=\pm\frac{\pi}{2}, and reaches its minimum, zero, at θji=0\theta_{ji}=0, θji=π\theta_{ji}=\pi, α=0\alpha=0, or α\alpha\to\infty. For a fixed θji0,±π2,π\theta_{ji}\neq 0,\pm\frac{\pi}{2},\pi, the minima occurs when α=0\alpha=0 or α\alpha\to\infty, but the maximum extreme point is not at α=1\alpha=1.

Next, we solve two assignment problems for selecting sensor(s) to track target(s).

IV Assignment Algorithms

So far, we have assumed that we know the true position, o(k)o(k), of the target at time kk. In practice, we only have an estimate, o^(k)\hat{o}(k), for oo along with its covariance Σ(k)\Sigma(k). The estimate is obtained by fusing the past measurements obtained by the sensors using, for example, an Extended Kalman Filter (EKF).

Since the sensors know the probability distribution of target’s state with mean o^\hat{o}, and covariance Σ\Sigma, they can calculate the relative distances, diod_{io}, by using the Mahalanobis distance [11],

dio=(o^pi)TΣ1(o^pi),i{1,,,N}.d_{io}=\sqrt{(\hat{o}-p_{i})^{T}\Sigma^{-1}(\hat{o}-p_{i})},~i\in\{1,,...,N\}.

IV-A A 1/31/3–approximation algorithm for Problem 1

We propose a greedy algorithm to solve Problem 1. In each round, we calculate the observability metric, ω(σ1(l),σ2(l),tl)\omega(\sigma_{1}(l),\sigma_{2}(l),t_{l}) for all triples (σ1(l),σ2(l),tl),σ1(l),σ2(l)𝒮,tl𝒯(\sigma_{1}(l),\sigma_{2}(l),t_{l}),~\sigma_{1}(l),\sigma_{2}(l)\in\mathcal{S},~t_{l}\in\mathcal{T}, and select the triple which has the maximum ω(σ1(l),σ2(l),tl)\omega(\sigma_{1}(l),\sigma_{2}(l),t_{l}), then remove {σ1(l),σ2(l)}\{\sigma_{1}(l),\sigma_{2}(l)\} from sensor set 𝒮\mathcal{S} and remove tlt_{l} from target set 𝒯\mathcal{T}, respectively. We present the greedy approach in Algorithm 1 where ω(GREEDY)\omega(\textrm{GREEDY}) denotes total value charged by the greedy approach. We can use the inverse of the condition number (Equation 9) as ω()\omega(\cdot).

k0,ω(GREEDY)0k\leftarrow 0,~\omega(\textrm{GREEDY})\leftarrow 0
while true do
   Compute all possible ω(σ1(l),σ2(l),tl)\omega(\sigma_{1}(l),\sigma_{2}(l),t_{l}).
   Select triple (σ1(l),σ2(l),tl)(\sigma_{1}(l),\sigma_{2}(l),t_{l}) with maximum ω(σ1(l),σ2(l),tl)\omega(\sigma_{1}(l),\sigma_{2}(l),t_{l}) defined as ωmax\omega_{\max}.
 ω(GREEDY)ω(GREEDY)+ωmax\omega(\textrm{GREEDY})\leftarrow\omega(\textrm{GREEDY})+\omega_{\max}.
   Remove {si,sj}\{s_{i},s_{j}\} from union sensor SS and remove tlt_{l} from target set 𝒯\mathcal{T}.
 kk+1k\leftarrow k+1
end while
Algorithm 1 Greedy Unique Pair Assignment

We present the following lemmas to guarantee the effectiveness of the greedy algorithm.

Theorem 6.

ω(GREEDY)13OPT\omega(\textrm{GREEDY})\geq\frac{1}{3}\textrm{OPT} where OPT is the optimal algorithm for Problem 1.

Proof.

Suppose GREEDY picks triple (si,sj,tl)(s_{i},s_{j},t_{l}) in the kthk^{th} round. We will charge the value ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}) to at most three triples in OPT that have not been charged in previous rounds. Furthermore, if a triple in OPT is charged in the kthk^{th} round, then we show that the ω()\omega(\cdot) value of the triple is less than ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}).

There are three cases.

  1. 1.

    (si,sj,tl)(s_{i},s_{j},t_{l}) is also chosen by OPT. In this case, we will charge exactly ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}) to the triple (si,sj,tl)(s_{i},s_{j},t_{l}) in OPT, if it has not been charged previously.

  2. 2.

    Exactly two of (si,sj,tl)(s_{i},s_{j},t_{l}) appear in a triple chosen by OPT. Consider the case where OPT chooses (si,sj,tm)(s_{i},s_{j},t_{m}) where mlm\neq l. All other cases are symmetric. In this case, we need to charge ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}) to at most two triples — (si,sj,tm)(s_{i},s_{j},t_{m}) and the one containing tlt_{l}, say (sp,sq,tl)(s_{p},s_{q},t_{l}). If (si,sj,tm)(s_{i},s_{j},t_{m}) has not been charged previously, then it must mean that GREEDY has not assigned any sensors to tmt_{m} in previous rounds. Since GREEDY chose (si,sj,tl)(s_{i},s_{j},t_{l}) in the kthk^{th} round, it must mean ω(si,sj,tl)ω(si,sj,tm)\omega(s_{i},s_{j},t_{l})\geq\omega(s_{i},s_{j},t_{m}), otherwise GREEDY would have chosen (si,sj,tm)(s_{i},s_{j},t_{m}) in the kthk^{th} round. Likewise, if (sp,sq,tl)(s_{p},s_{q},t_{l}) has not been charged previously, we must have ω(si,sj,tl)ω(sp,sq,tl)\omega(s_{i},s_{j},t_{l})\geq\omega(s_{p},s_{q},t_{l}). Therefore, the value charged in the kthk^{th} round will be at most twice of (si,sj,tl)(s_{i},s_{j},t_{l}).

  3. 3.

    No two of (si,sj,tl)(s_{i},s_{j},t_{l}) appear in the same triple chosen by OPT. In this case, we can charge the value of ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}) to at most three triples. Using an argument similar to the previous case, we can say that the value charged in the kthk^{th} round will be at most thrice of ω(si,sj,tl)\omega(s_{i},s_{j},t_{l}).

Therefore, once all triples in GREEDY are charged, it follows that ω(GREEDY)13OPT\omega(\textrm{GREEDY})\geq\frac{1}{3}\textrm{OPT}.

IV-B General assignment using Submodular Welfare Optimization

The sensor assignment problem where only two sensors are assigned is discussed above. We now study a more general assignment (Problem 2) where each target tlt_{l} is tracked by a subset of sensors σ(l)S,l{1,2,,L}\sigma(l)\subset S,~l\in\{1,2,...,L\} whose cardinality is not necessarily two. This is known as submodular welfare problem in the literature[12] where the objective is to maximize i=1nwi(Si)\sum_{i=1}^{n}w_{i}(S_{i}) for independent sets {Si|SiS,i={1,2,n}}\{S_{i}|S_{i}\in S,~i=\{1,2,...n\}\} by using monotone and submodular utility functions wiw_{i}. A greedy algorithm [9] yields a 1/21/2–approximation for this problem. We first show that the lower bound of the inverse of the condition number is neither monotone nor submodular which makes the optimization problem much more challenging.

Theorem 7.

The lower bound of the inverse of condition number function ω()\omega(\cdot) is neither monotone increasing nor submodular.

Proof.

We prove the claim by giving two counter-examples.

Case 1: Given the sensors s1(0,0)s_{1}(0,0), s2(23,9)s_{2}(2\sqrt{3},-9), s3(3,3)s_{3}(\sqrt{3},3) and target t1(3,1)t_{1}(\sqrt{3},1) with uo1,max=1u_{o_{1},max}=1 in 2-D plane (Figure 3-(a)), ω({s1,s3})=0.5345>ω({s1,s2,s3})=0.1823\omega(\{s_{1},s_{3}\})=0.5345>\omega(\{s_{1},s_{2},s_{3}\})=0.1823, which shows ω()\omega(\cdot) is not monotone increasing.

Case 2: Given the sensors s1(0,0)s_{1}(0,0), s2(23,0)s_{2}(2\sqrt{3},0), s3(3,0.1)s_{3}(\sqrt{3},0.1), s4(3,3)s_{4}(\sqrt{3},3) and target t1(3,1)t_{1}(\sqrt{3},1) with uo1,max=1u_{o_{1},max}=1 in 2-D plane (Figure 3-(b)), ω({s1,s2,s3})ω({s1,s2})=0.33100.5345=0.2035<ω({s1,s2,s4,s3})ω({s1,s2,s4})=0.87650.9258=0.0493\omega(\{s_{1},s_{2},s_{3}\})-\omega(\{s_{1},s_{2}\})=0.3310-0.5345=-0.2035<\omega(\{s_{1},s_{2},s_{4},s_{3}\})-\omega(\{s_{1},s_{2},s_{4}\})=0.8765-0.9258=-0.0493, which shows ω()\omega(\cdot) is not submodular.

Refer to caption
(a) Case 1
Refer to caption
(b) Case 2
Figure 3: Positions of sensors and target in 2D plane.

Therefore, we focus on other measures of observability and summarize the results in Theorem 8.

Theorem 8.

For symmetric observability matrix, 𝕆(o,uo):=OT(o,uo)O(o,uo)\mathbb{O}(o,u_{o}):=O^{T}(o,u_{o})O(o,u_{o}), its trace, log determinant, rank and the trace of its inverse are submodular and monotone increasing.

The proof is similar to proving that the trace of the Gramian and inverse Gramian, the log determinant, and the rank of the Gramian are monotone submodular[13]. Resorting to submodular and monotone observability measure, we can use a simpler greedy algorithm to solve the General Assignment Problem.

V Simulations

We illustrate the performance of the pairing and assignment strategies for sensor selection using observability measure as the performance criterion. We first consider the two sensors case for tracking a moving target, and then focus on the multi-sensor multi-target assignment problems. The video444https://youtu.be/Kt0yeYXyrKQ and the code555https://github.com/lovetuliper/observability-based-sensor-assignment.git of our simulations are available online.

V-A Two Sensors Case

Refer to caption
(a) k=1k=1
Refer to caption
(b) k=50k=50
Refer to caption
(c) k=150k=150
Refer to caption
(d) k=200k=200
Refer to caption
(e) k=250k=250
Refer to caption
(f) k=315k=315
Figure 4: Strategy (1) in action for tracking the target with a circular motion.
Refer to caption
Refer to caption
Figure 5: (a) Best pair selection at each timestep. (b) Best partner selection for sensor s2s_{2} at each timestep.

We first consider the scenario where the target is moving independently of six stationary sensors (s1s_{1}s6s_{6}). The target follows a circular trajectory with uo,max=3u_{o,\max}=3. We compare three strategies: (1) flexible best pair: we pair sensors (si,sj)(s_{i},s_{j}) which have the maximum C¯1(Oi,j(o^(k1),uo,max))\underline{C}^{-1}(O_{i,j}(\hat{o}(k-1),u_{o,\max})) for tracking the moving target oo (Figure 5–(a)); (2) flexible pair for fixed sensor: we pair an arbitrarily chosen sensor, s2s_{2}, with the best sensor at each timestep (Figure 5–(b)); (3) best-fixed pair: we pair s2s_{2} with s1s_{1} at all timesteps. We chose s1s_{1} since it gives the best performance amongst all other sensors, on an average, for the circular trajectory.

Figure 4–(a) to (f) shows the result of running strategy (1) over 315315 timesteps. The best selected pair at each timestep kk is shown in Figure 5-(a).

We show the results for the three strategies in Figure 6. We plot the inverse of the condition number, the estimation error ϵ:=o^o2\epsilon:=\|\hat{o}-o\|_{2}, and the trace of covariance matrix over time. We observe that the first strategy maintains a higher C¯1(Oi,j(o^,uo,max))\underline{C}^{-1}(O_{i,j}(\hat{o},u_{o,\max})) and has lower ϵ\epsilon and trace of covariance almost all times. Around k=50k=50, the best pair changes frequently as seen in Figure 5-(a).

Refer to caption
(a) flexible best pairing
Refer to caption
(b) best pairing for a fixed sensor s2s_{2}
Refer to caption
(c) best fixed partner, s1s_{1}, for s2s_{2}
Figure 6: Comparison of lower bound of the inverse of condition number, estimate error, and the trace of covariance matrix in three scenarios.

Note that in all three cases, higher C¯1(Oi,j(o^,uo,max))\underline{C}^{-1}(O_{i,j}(\hat{o},u_{o,\max})) correlates with lower ϵ\epsilon and trace(Σ)\mathrm{trace}(\Sigma) suggesting that the tracking performance can be improved by improving the lower bound of the inverse of the condition number.

V-B Adversarial Target

We also simulate the scenario where the target moves in an adversarial fashion. At each time step, the target moves in a direction so as to minimize the inverse of the condition number. Here, the target has an advantage in that it knows its own exact state and control input, while sensors only have an estimate. At each timestep, the target evaluates all the possible control inputs within a ball of radius uo,maxu_{o,\max}, around itself and chooses on that minimizes C1(Oi,j(o,uo)){C}^{-1}(O_{i,j}(o,u_{o})).

Figure 7 shows an instance where the target is “trapped” by the sensors. The target fails to reduce the inverse of the condition number since the sensor pair switches whenever the target gets closer to a sensor.

Refer to caption
Refer to caption
Figure 7: An adversarial target is “entrapped” by sensors using best pairing strategy. The inverse of the condition number remains at a high value. Even if the target moves closer to a sensor to reduce the inverse of the condition number, the sensor pair switches thereby increasing the inverse of the condition number.

V-C Greedy Unique Pair Assignment

The Unique Pair Assignment problem is NP-Complete. Therefore, finding ω(OPT)\omega(\textrm{OPT}) is infeasible in polynomial time. In order to empirically evaluate the Greedy Unique Pair Assignment (Algorithm 1), we resort to a new assignment scenario where each sensor can be matched in multiple different pairs and each sensor pair can be assigned to at most one specific target. We formulate the new assignment as Relaxed Pair Assignment (Problem 3). It is clear that solving Relaxed Pair Assignment problem optimally gives us an upper bound of optimality for Unique Pair Assignment problem. We can use this upper bound for the comparison of the greedy approach in Unique Pair Assignment.

Problem 3 (Relaxed Pair Assignment).

Given a set of sensor positions, 𝒮:={s0,,sN}\mathcal{S}:=\{s_{0},\ldots,s_{N}\} and a set of target estimates at time tt, 𝒯:={t0,,tL}\mathcal{T}:=\{t_{0},\ldots,t_{L}\}, find an assignment of unique pairs of sensors to targets:

maximizel=1Lω(σ1(l),σ2(l),tl)\text{maximize}\sum_{l=1}^{L}\omega(\sigma_{1}(l),\sigma_{2}(l),t_{l}) (10)

with the added constraint that all pairs are unique, that is, k,l=1,,L\forall k,l=1,\ldots,L, klk\neq l, σ1(k)σ1(l)\sigma_{1}(k)\neq\sigma_{1}(l) and/or σ2(k)σ2(l)\sigma_{2}(k)\neq\sigma_{2}(l).

The Relaxed Pair Assignment problem can be solved optimally by using maximum weight perfect bipartite matching (MWPBM) [14]. Note that a sensor can be matched in multiple distinct pairs and assigned to multiple targets. The MWPBM can be solved using the Hungarian algorithm [15] in polynomial time.

In order to see the effectiveness of the greedy algorithm in Unique Pair Assignment, we compare the total value charged by the greedy algorithm, ω(GREEDY)\omega(\textrm{GREEDY}), with the total value charged by the MWPBM, ω(MWPBM)\omega(\textrm{MWPBM}), in Relaxed Pair Assignment as shown in Figure 8. We consider different number of targets with LL for 1 to 20 and set the number of robots, N=2LN=2L. For each L{1,2,,20}L\in\{1,2,...,20\}, the positions of sensors and targets are randomly generated within [0,100]×[0,100]2[0,100]\times[0,100]\in\mathbb{R}^{2} for 30 trials. Set the maximum control input for each target as uo,max=1u_{o,\max}=1. Figure 8 shows that ω(GREEDY)\omega(\textrm{GREEDY}) is close to ω(MWPBM)\omega(\textrm{MWPBM}) and much higher than 13ω(MWPBM)\frac{1}{3}\omega(\textrm{MWPBM}). Thus, even though we give a theoretical 1/31/3–approximation for the greedy algorithm, it performs much better in practice.

Refer to caption
Figure 8: Comparison of total reward collected by greedy approach (Algorithm 1) with the maximum perfect pair matching.

V-D General Pair Assignment

We also simulate the greedy assignment [9] for the General Assignment problem by using log determinant of symmetric observability matrix, 𝕆(o,uo)\mathbb{O}(o,u_{o}) as observability measure. We set the number of targets as L=5L=5 and number of sensors from 20 to 50. For each N{20,21,,50}N\in\{20,21,...,50\}, the positions of sensors and targets are randomly generated within [0,100]×[0,100]2[0,100]\times[0,100]\in\mathbb{R}^{2} for 30 trials. We compare the number of sensors assigned to one specific target, i.e., tl,l{1,2,L}t_{l},l\in\{1,2,...L\} with the N/LN/L as shown in Figure 9. It shows that the sensors are assigned to each target almost evenly.

Refer to caption
Figure 9: Number of sensors assigned to each target.

VI Conclusion

In this paper, we solved sensor assignment problems to improve the observability for target tracking. We derived the lower bound on the inverse of the condition number of the observability matrix for a system with a mobile target and NN stationary sensors. The lower bound considers only the known part of the observability matrix — the sensor-target relative position and an upper bound on the target’s speed. We showed how this lower bound can be employed for sensor selection. We considered two sensor assignment problems for which we presented constant-factor approximation algorithm. Our immediate work is focused on assigning sensors to cover an area instead of tracking a group of targets. Another avenue is designing an efficient set covering strategy based on observability measures which are submodular and monotone.

References

  • [1] M. Montemerlo, S. Thrun, D. Koller, B. Wegbreit, et al., “Fastslam: A factored solution to the simultaneous localization and mapping problem,” in Aaai/iaai, 2002, pp. 593–598.
  • [2] H. Durrant-Whyte and T. Bailey, “Simultaneous localization and mapping: part i,” IEEE robotics & automation magazine, vol. 13, no. 2, pp. 99–110, 2006.
  • [3] A. S. Gadre and D. J. Stilwell, “Toward underwater navigation based on range measurements from a single location,” in Robotics and Automation, 2004. Proceedings. ICRA’04. 2004 IEEE International Conference on, vol. 5. IEEE, 2004, pp. 4472–4477.
  • [4] G. Papadopoulos, M. F. Fallon, J. J. Leonard, and N. M. Patrikalakis, “Cooperative localization of marine vehicles using nonlinear state estimation,” in Intelligent Robots and Systems (IROS), 2010 IEEE/RSJ International Conference on. IEEE, 2010, pp. 4874–4879.
  • [5] F. Arrichiello, G. Antonelli, A. P. Aguiar, and A. Pascoal, “An observability metric for underwater vehicle localization using range measurements,” Sensors, vol. 13, no. 12, pp. 16 191–16 215, 2013.
  • [6] R. K. Williams and G. S. Sukhatme, “Observability in topology-constrained multi-robot target tracking,” in Robotics and Automation (ICRA), 2015 IEEE International Conference on. IEEE, 2015, pp. 1795–1801.
  • [7] R. Hermann and A. Krener, “Nonlinear controllability and observability,” IEEE Transactions on automatic control, vol. 22, no. 5, pp. 728–740, 1977.
  • [8] A. J. Krener and K. Ide, “Measures of unobservability,” in Decision and Control, 2009 held jointly with the 2009 28th Chinese Control Conference. CDC/CCC 2009. Proceedings of the 48th IEEE Conference on. IEEE, 2009, pp. 6401–6406.
  • [9] G. L. Nemhauser, L. A. Wolsey, and M. L. Fisher, “An analysis of approximations for maximizing submodular set functions—i,” Mathematical Programming, vol. 14, no. 1, pp. 265–294, 1978.
  • [10] P. Tokekar, J. Vander Hook, and V. Isler, “Active target localization for bearing based robotic telemetry,” in Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2011, pp. 488–493.
  • [11] P. C. Mahalanobis, “On the generalized distance in statistics,” Proceedings of the National Institute of Sciences (Calcutta), vol. 2, pp. 49–55, 1936.
  • [12] J. Vondrák, “Optimal approximation for the submodular welfare problem in the value oracle model,” in Proceedings of the fortieth annual ACM symposium on Theory of computing. ACM, 2008, pp. 67–74.
  • [13] T. H. Summers, F. L. Cortesi, and J. Lygeros, “On submodularity and controllability in complex dynamical networks,” IEEE Transactions on Control of Network Systems, vol. 3, no. 1, pp. 91–101, 2016.
  • [14] T. H. Cormen, Introduction to algorithms. MIT press, 2009.
  • [15] H. W. Kuhn, “The hungarian method for the assignment problem,” Naval research logistics quarterly, vol. 2, no. 1-2, pp. 83–97, 1955.
  • [16] G. Strang, Linear algebra and its applications. Academic Press, 1976.
  • [17] J. N. Franklin, Matrix theory. Courier Corporation, 2012.

Proof for Theorem 1

Proof.

We partition O(o,uo)O(o,u_{o}) into the known and unknown parts as,

O(o,uo)=[O(o)O(uo)],\displaystyle O(o,u_{o})=\begin{bmatrix}O(o)\\ O(u_{o})\end{bmatrix}, (11)

where

O(o):=[oxp1x,oyp1yoxp2x,oyp2yoxpNx,oypNy],O(o):=\begin{bmatrix}o_{x}-p_{1x},o_{y}-p_{1y}\\ o_{x}-p_{2x},o_{y}-p_{2y}\\ \vdots\\ o_{x}-p_{Nx},o_{y}-p_{Ny}\end{bmatrix}, (12)

and

O(uo):=[uox,uoy],O(u_{o}):=\begin{bmatrix}u_{ox},u_{oy}\end{bmatrix}, (13)

indicate the contribution to the observability matrix from sensor-target relative state and target’s control input, respectively.

The singular values of O(o,uo)O(o,u_{o}) can be found as the square-root of the eigenvalues of the symmetric observability matrix, 𝕆(o,uo)\mathbb{O}(o,u_{o}), given as [16],

𝕆(o,uo)\displaystyle\mathbb{O}(o,u_{o}) =\displaystyle= OT(o,uo)O(o,uo)\displaystyle O^{T}(o,u_{o})O(o,u_{o}) (14)
=\displaystyle= OT(o)O(o)+OT(uo)O(uo).\displaystyle O^{T}(o)O(o)+O^{T}(u_{o})O(u_{o}).
{λmin(𝕆(o,uo))=σmin(O(o,uo)),λmax(𝕆(o,uo))=σmax(O(o,uo)).\left\{\begin{array}[]{ll}\sqrt{{\lambda_{\min}(\mathbb{O}(o,u_{o}))}}={\sigma_{\min}(O(o,u_{o}))},\\ \sqrt{{\lambda_{\max}(\mathbb{O}(o,u_{o}))}}={\sigma_{\max}(O(o,u_{o}))}.\end{array}\right. (15)

We can use Weyl and dual Weyl inequalities to bound the singular values. For Hermitian matrices XX and YY with rr eigenvalues written in increasing order λ1(X)λ2(X)λr(X)\lambda_{1}(X)\leq\lambda_{2}(X)\leq\ldots\leq\lambda_{r}(X) and λ1(Y)λ2(Y)λr(Y)\lambda_{1}(Y)\leq\lambda_{2}(Y)\leq\ldots\leq\lambda_{r}(Y), respectively, the Weyl inequalities[17] is given by,

λi+j1(X+Y)λi(X)+λj(Y)\lambda_{i+j-1}(X+Y)\geq\lambda_{i}(X)+\lambda_{j}(Y) (16)

where i,j1i,j\geq 1 and i+j1ri+j-1\leq r. Similarly, the dual Weyl inequalities is given by

λi+jr(X+Y)λi(X)+λj(Y)\lambda_{i+j-r}(X+Y)\leq\lambda_{i}(X)+\lambda_{j}(Y) (17)

where i,j1i,j\geq 1 and i+jrri+j-r\leq r.

Since 𝕆(o,uo)2×2\mathbb{O}(o,u_{o})\in\mathbb{R}^{2\times 2}, OT(o)O(o)2×2O^{T}(o)O(o)\in\mathbb{R}^{2\times 2} and OT(uo)O(uo)2×2O^{T}(u_{o})O(u_{o})\in\mathbb{R}^{2\times 2} are symmetric matrices, they are Hermitian with the eigenvalues (in ascending order) as λ1(𝕆(o,uo))λ2(𝕆(o,uo))\lambda_{1}(\mathbb{O}(o,u_{o}))\leq\lambda_{2}(\mathbb{O}(o,u_{o})), λ1(OT(o)O(o))λ2(OT(o)O(o))\lambda_{1}(O^{T}(o)O(o))\leq\lambda_{2}(O^{T}(o)O(o)) and λ1(OT(uo)O(uo))λ2(OT(uo)O(uo))\lambda_{1}(O^{T}(u_{o})O(u_{o}))\leq\lambda_{2}(O^{T}(u_{o})O(u_{o})). Following the Weyl and dual Weyl inequalities, we get

λ1(𝕆(o,uo))λ1(OT(o)O(o))+λ1(OT(uo)O(uo)),\displaystyle\lambda_{1}(\mathbb{O}(o,u_{o}))\geq\lambda_{1}(O^{T}(o)O(o))+\lambda_{1}(O^{T}(u_{o})O(u_{o})),
λ1(𝕆(o,uo))\displaystyle\lambda_{1}(\mathbb{O}(o,u_{o}))\leq
min{λ1(OT(o)O(o))+λ2(OT(uo)O(uo))λ2(OT(o)O(o))+λ1(OT(uo)O(uo)),\displaystyle\min\left\{\begin{array}[]{ll}\lambda_{1}(O^{T}(o)O(o))+\lambda_{2}(O^{T}(u_{o})O(u_{o}))\\ \lambda_{2}(O^{T}(o)O(o))+\lambda_{1}(O^{T}(u_{o})O(u_{o}))\end{array},\right. (20)
λ2(𝕆(o,uo))λ2(OT(o)O(o))+λ2(OT(uo)O(uo)),\displaystyle\lambda_{2}(\mathbb{O}(o,u_{o}))\leq\lambda_{2}(O^{T}(o)O(o))+\lambda_{2}(O^{T}(u_{o})O(u_{o})),
λ2(𝕆(o,uo))\displaystyle\lambda_{2}(\mathbb{O}(o,u_{o}))\geq
max{λ1(OT(o)O(o))+λ2(OT(uo)O(uo))λ2(OT(o)O(o))+λ1(OT(uo)O(uo)).\displaystyle\max\left\{\begin{array}[]{ll}\lambda_{1}(O^{T}(o)O(o))+\lambda_{2}(O^{T}(u_{o})O(u_{o}))\\ \lambda_{2}(O^{T}(o)O(o))+\lambda_{1}(O^{T}(u_{o})O(u_{o}))\end{array}.\right. (23)

Thus,

λ1(𝕆(o,uo))λ2(𝕆(o,uo))\displaystyle\frac{\lambda_{1}(\mathbb{O}(o,u_{o}))}{\lambda_{2}(\mathbb{O}(o,u_{o}))}\geq
λ1(OT(o)O(o))+λ1(OT(uo)O(uo))λ2(OT(o)O(o))+λ2(OT(uo)O(uo)).\displaystyle\frac{\lambda_{1}(O^{T}(o)O(o))+\lambda_{1}(O^{T}(u_{o})O(u_{o}))}{\lambda_{2}(O^{T}(o)O(o))+\lambda_{2}(O^{T}(u_{o})O(u_{o}))}. (24)

Then, from Equation 15 and Equation 24, the inverse of the condition number of the local nonlinear observability matrix,

C1(O(o,uo))=λ1(𝕆(o,uo))λ2(𝕆(o,uo))\displaystyle C^{-1}(O(o,u_{o}))=\sqrt{\frac{\lambda_{1}(\mathbb{O}(o,u_{o}))}{\lambda_{2}(\mathbb{O}(o,u_{o}))}}
λ1(OT(o)O(o))+λ1(OT(uo)O(uo))λ2(OT(o)O(o))+λ2(OT(uo)O(uo)).\displaystyle\geq\sqrt{\frac{\lambda_{1}(O^{T}(o)O(o))+\lambda_{1}(O^{T}(u_{o})O(u_{o}))}{\lambda_{2}(O^{T}(o)O(o))+\lambda_{2}(O^{T}(u_{o})O(u_{o}))}}.

By calculating the eigenvalues of symmetric matrix of target’s control contribution,

OT(uo)O(uo)=[uox2uoxuoyuoyuoxuoy2],O^{T}(u_{o})O(u_{o})=\begin{bmatrix}u_{ox}^{2}&u_{ox}u_{oy}\\ u_{oy}u_{ox}&u_{oy}^{2}\end{bmatrix},

we get,

λ1(OT(uo)O(uo))=0\displaystyle\lambda_{1}(O^{T}(u_{o})O(u_{o}))=0
λ2(OT(uo)O(uo))=uox2+uoy2=uo2.\displaystyle\lambda_{2}(O^{T}(u_{o})O(u_{o}))=u_{ox}^{2}+u_{oy}^{2}=u_{o}^{2}. (25)

Then the lower bound of C1(O(o,uo))C^{-1}(O(o,u_{o})) is calculated as

C¯1(O(o,uo))\displaystyle\underline{C}^{-1}(O(o,u_{o})) =\displaystyle= λ1(OT(o)O(o))λ2(OT(o)O(o))+uo2\displaystyle\sqrt{\frac{\lambda_{1}(O^{T}(o)O(o))}{\lambda_{2}(O^{T}(o)O(o))+u_{o}^{2}}} (26)
=\displaystyle= σmin(O(o))σmax2(O(o))+uo2\displaystyle\frac{\sigma_{\min}(O(o))}{\sqrt{\sigma^{2}_{\max}(O(o))+u_{o}^{2}}}

Equation 26 gives the main lower bound. Note that C1(O(o,uo))C^{-1}(O(o,u_{o})) cannot be determined since target’s control input, uou_{o}, is unknown. However, we know that uo2uo,max||u_{o}||_{2}\leq u_{o,\max}. Therefore,

C¯1(O(o,uo))σmin(O(o))σmax2(O(o))+uo,max2\underline{C}^{-1}(O(o,u_{o}))\geq\frac{\sigma_{\min}(O(o))}{\sqrt{\sigma^{2}_{\max}(O(o))+u_{o,\max}^{2}}} (27)

This yields our main lower bound result.

Proof for Theorem 2

Proof.

The local observability matrix for one-sensor-target, ioi-o system can be derived from Equation 5 as,

Oi(o,uo)=[oxpixoypiyuoxuoy].O_{i}(o,u_{o})=\begin{bmatrix}o_{x}-p_{ix}&o_{y}-p_{iy}\\ u_{ox}&u_{oy}\end{bmatrix}. (28)

The sensor-target relative state contribution is.

Oi(o)=[oxpixoypiy].O_{i}(o)=\begin{bmatrix}o_{x}-p_{ix}&o_{y}-p_{iy}\end{bmatrix}.

The ioi-o system is weakly locally observable if Oi(o,uo)O_{i}(o,u_{o}) has full column rank, i.e., (oxpix)uoy(oypiy)uox(o_{x}-p_{ix})u_{oy}\neq(o_{y}-p_{iy})u_{ox}. However, the sensor does not know the target’s control input, uou_{o}.

From the eigenvalues of symmetric matrix of sensor-target relative state contribution of ioi-o system given by,

OiT(o)Oi(o)=[(oxpix)2,(oxpix)(oypiy)(oxpix)(oypiy),(oypiy)2],O_{i}^{T}(o)O_{i}(o)=\begin{bmatrix}(o_{x}-p_{ix})^{2},(o_{x}-p_{ix})(o_{y}-p_{iy})\\ (o_{x}-p_{ix})(o_{y}-p_{iy}),(o_{y}-p_{iy})^{2}\end{bmatrix},

we get,

{σmin(Oi(o))=λmin(OiT(o)Oi(o))=0,σmax(Oi(o))=λmax(OiT(o)Oi(o))=(oxpix)2+(oypiy)2.\left\{\begin{array}[]{ll}{\sigma_{\min}(O_{i}(o))}=\sqrt{{\lambda_{\min}(O_{i}^{T}(o)O_{i}(o))}}=0,\\ {\sigma_{\max}(O_{i}(o))}=\sqrt{{\lambda_{\max}(O_{i}^{T}(o)O_{i}(o))}}\\ ~~~~~~~~~~~~~~~=\sqrt{(o_{x}-p_{ix})^{2}+(o_{y}-p_{iy})^{2}}.\end{array}\right. (29)

Thus, from Equation 26, the lower bound for C1(Oi(o,uo))C^{-1}(O_{i}(o,u_{o})) is σmin(Oi(o))σmax2(Oi(o))+uo2=0\frac{\sigma_{\min}(O_{i}(o))}{\sqrt{\sigma^{2}_{\max}(O_{i}(o))+u_{o}^{2}}}=0. Consequently, the lower bound cannot be controlled by the sensor.

Proof for Theorem 3

Proof.

Recall that σmin(O(o,uo))=λmin(𝕆(o,uo)){\sigma_{\min}(O(o,u_{o}))}=\sqrt{{\lambda_{\min}(\mathbb{O}(o,u_{o}))}}. We have,

σmin(O(o,uo))σmax(O(o,uo))=λmin(𝕆(o,uo))λmax(𝕆(o,uo)).\frac{\sigma_{\min}(O(o,u_{o}))}{\sigma_{\max}(O(o,u_{o}))}=\frac{\sqrt{\lambda_{\min}(\mathbb{O}(o,u_{o}))}}{\sqrt{\lambda_{\max}(\mathbb{O}(o,u_{o}))}}. (30)

Following Equations 24 and 25, we have the lower bound for λmin(𝕆(o,uo))λmax(𝕆(o,uo))\frac{\lambda_{\min}(\mathbb{O}(o,u_{o}))}{\lambda_{\max}(\mathbb{O}(o,u_{o}))}, described as

λmin(𝕆(o,uo))λmax(𝕆(o,uo))λmin(OT(o)O(o))λmax(OT(o)O(o))+uo2,\displaystyle\frac{\lambda_{\min}(\mathbb{O}(o,u_{o}))}{\lambda_{\max}(\mathbb{O}(o,u_{o}))}\geq\frac{\lambda_{\min}(O^{T}(o)O(o))}{\lambda_{\max}(O^{T}(o)O(o))+u_{o}^{2}}, (31)

The observability can be improved by increasing this lower bound. We can transform the statement of the theorem (using eigenvalues instead of singular values) as: if

{λmin(OT(o)O(o))λmax(OT(o)O(o))λmin(OT(o)O(o))λmax(OT(o)O(o)),λmin(OT(o)O(o))λmax(OT(o)O(o))\left\{\begin{array}[]{ll}\frac{\lambda_{\min}^{{}^{\prime}}(O^{T}(o)O(o))}{\lambda_{\max}^{{}^{\prime}}(O^{T}(o)O(o))}\geq\frac{\lambda_{\min}(O^{T}(o)O(o))}{\lambda_{\max}(O^{T}(o)O(o))},\\ \lambda_{\min}^{{}^{\prime}}(O^{T}(o)O(o))\geq\lambda_{\max}(O^{T}(o)O(o))\end{array}\right. (32)

then

λmin(OT(o)O(o))λmax(OT(o)O(o))+uo2\displaystyle\frac{\lambda_{\min}^{{}^{\prime}}(O^{T}(o)O(o))}{\lambda_{\max}^{{}^{\prime}}(O^{T}(o)O(o))+u_{o}^{2}}
λmin(OT(o)O(o))λmax(OT(o)O(o))+uo2\displaystyle\geq\frac{\lambda_{\min}(O^{T}(o)O(o))}{\lambda_{\max}(O^{T}(o)O(o))+u_{o}^{2}} (33)

where the λ\lambda and λ\lambda^{\prime} denotes the eigenvalues before and after the sensors’ apply their control, respectively.

We start with the left-hand side of Equation 33) to get,

λmin,oλmax,o+uo2λmin,oλmax,o+uo2\displaystyle\frac{\lambda_{\min,o}^{{}^{\prime}}}{\lambda_{\max,o}^{{}^{\prime}}+u_{o}^{2}}-\frac{\lambda_{\min,o}}{\lambda_{\max,o}+u_{o}^{2}} (34)
=\displaystyle= λmin,oλmax,oλmax,oλmin,o(λmax,o+uo2)(λmax,o+uo2)\displaystyle\frac{\lambda_{\min,o}^{{}^{\prime}}\lambda_{\max,o}-\lambda_{\max,o}^{{}^{\prime}}\lambda_{\min,o}}{(\lambda_{\max,o}^{{}^{\prime}}+u_{o}^{2})(\lambda_{\max,o}+u_{o}^{2})}
+\displaystyle+ uo2(λmin,oλmin,o)(λmax,o+uo2)(λmax,o+uo2)\displaystyle\frac{u_{o}^{2}(\lambda_{\min,o}^{{}^{\prime}}-\lambda_{\min,o})}{(\lambda_{\max,o}^{{}^{\prime}}+u_{o}^{2})(\lambda_{\max,o}+u_{o}^{2})}

where λmin,o\lambda_{\min,o}^{{}^{\prime}}, λmax,o\lambda_{\max,o}^{{}^{\prime}}, λmin,o\lambda_{\min,o} and λmax,o\lambda_{\max,o} indicate λmin(OT(o)O(o))\lambda_{\min}^{{}^{\prime}}(O^{T}(o)O(o)), λmax(OT(o)O(o))\lambda_{\max}^{{}^{\prime}}(O^{T}(o)O(o)), λmin(OT(o)O(o))\lambda_{\min}(O^{T}(o)O(o)) and λmax(OT(o)O(o))\lambda_{\max}(O^{T}(o)O(o)).

Using Equation 32, it is easy to show

λmin,oλmax,o+uo2λmin,oλmax,o+uo20.\frac{\lambda_{\min,o}^{{}^{\prime}}}{\lambda_{\max,o}^{{}^{\prime}}+u_{o}^{2}}-\frac{\lambda_{\min,o}}{\lambda_{\max,o}+u_{o}^{2}}\geq 0. (35)

Hence the claim is proved.

Remark 2.

Equation 32 is sufficient, but not necessary condition to guarantee Equation 33. This is because Equation 33 can be established with a weaker condition,

λmin,oλmax,oλmax,oλmin,o+uo2(λmin,oλmin,o)0.\lambda_{\min,o}^{{}^{\prime}}\lambda_{\max,o}-\lambda_{\max,o}^{{}^{\prime}}\lambda_{\min,o}+u_{o}^{2}(\lambda_{\min,o}^{{}^{\prime}}-\lambda_{\min,o})\geq 0.

We choose the stricter condition (Equation 32) because it is conceptually easy to separate and eliminate the influence on the degree of observability from target’s control input, uou_{o}, which is unknown and uncontrolled.

Proof for Theorem 4

Proof.

The sensor-target relative state contribution of the local observability matrix Oi,j(o,uo)O_{i,j}(o,u_{o}) (Equation 7) is,

Oi,j(o)=[oxpixoypiyoxpjxoypjy].O_{i,j}(o)=\begin{bmatrix}o_{x}-p_{ix}&o_{y}-p_{iy}\\ o_{x}-p_{jx}&o_{y}-p_{jy}\end{bmatrix}. (36)

The lower bound of the inverse of the condition number of Oi,j(o,uo)O_{i,j}(o,u_{o}) is C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})). Our goal is to determine how to improve C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) by analyzing the sensor-target relative state contribution. This is similar to the work presented in [5].

We start by calculating the eigenvalues of Oi,jT(o)Oi,j(o)O_{i,j}^{T}(o)O_{i,j}(o).

Oi,jT(o)Oi,j(o)=[(oxpix)2,(oxpjx)(oypiy)(oypiy)(oxpjx),(oypjy)2].O_{i,j}^{T}(o)O_{i,j}(o)=\begin{bmatrix}(o_{x}-p_{ix})^{2},(o_{x}-p_{jx})(o_{y}-p_{iy})\\ (o_{y}-p_{iy})(o_{x}-p_{jx}),(o_{y}-p_{jy})^{2}\end{bmatrix}. (37)

Using polar coordinates (Equation 8), Oi,jT(o)Oi,j(o)O_{i,j}^{T}(o)O_{i,j}(o) can be rewritten as,

[dio2cos2θi+djo2cos2θj,dio2sinθicosθi+djo2sinθjcosθjdjo2sinθjcosθj+dio2sinθicosθi,dio2sin2θi+djo2sin2θj].\displaystyle\begin{bmatrix}d_{io}^{2}\cos^{2}\theta_{i}$+$d_{jo}^{2}\cos^{2}\theta_{j},d_{io}^{2}\sin\theta_{i}\cos\theta_{i}$+$d_{jo}^{2}\sin\theta_{j}\cos\theta_{j}\\ d_{jo}^{2}\sin\theta_{j}\cos\theta_{j}$+$d_{io}^{2}\sin\theta_{i}\cos\theta_{i},d_{io}^{2}\sin^{2}\theta_{i}$+$d_{jo}^{2}\sin^{2}\theta_{j}\end{bmatrix}.

The eigenvalues are given by,

λmin(Oi,jT(o)Oi,j(o))=\displaystyle\lambda_{\min}(O_{i,j}^{T}(o)O_{i,j}(o))=
dio2+djo2dio4+djo4+2dio2djo2cos(2θji)2,\displaystyle\frac{d_{io}^{2}+d_{jo}^{2}-\sqrt{d_{io}^{4}+d_{jo}^{4}+2d_{io}^{2}d_{jo}^{2}\cos(2\theta_{ji})}}{2},
λmax(Oi,jT(o)Oi,j(o))=\displaystyle\lambda_{\max}(O_{i,j}^{T}(o)O_{i,j}(o))=
dio2+djo2+dio4+djo4+2dio2djo2cos(2θji)2.\displaystyle\frac{d_{io}^{2}+d_{jo}^{2}+\sqrt{d_{io}^{4}+d_{jo}^{4}+2d_{io}^{2}d_{jo}^{2}\cos(2\theta_{ji})}}{2}. (38)

where θji:=θjθi\theta_{ji}:=\theta_{j}-\theta_{i}. Then, by Equation  26 in the proof of Theorem 1, the claim is guaranteed.

Proof for Theorem 5

Proof.

First, for a fixed α\alpha,

C¯1(Oi,j(o,uo))θji=0α2sin(2θji)=0,\frac{\partial\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\theta_{ji}}=0\Rightarrow\alpha^{2}\sin(2\theta_{ji})=0,
θji=0,±π2,π.\Rightarrow\theta_{ji}=0,\pm\frac{\pi}{2},\pi.

and

2C¯1(Oi,j(o,uo))θji2|θji=0,π>0.\frac{\partial^{2}\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\theta_{ji}^{2}}|_{\theta_{ji}=0,\pi}>0.
2C¯1(Oi,j(o,uo))θji2|θji=±π2<0.\frac{\partial^{2}\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\theta_{ji}^{2}}|_{\theta_{ji}=\pm\frac{\pi}{2}}<0.

Thus, when α\alpha is fixed, C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) reaches its minimum, C¯1(Oi,j(o,uo))|θji=0,π=0\underline{C}^{-1}(O_{i,j}(o,u_{o}))|_{\theta_{ji}=0,~\pi}=0, and reaches its maximum C¯1(Oi,j(o,uo))|θji=±π2=1+α2|1α2|1+α2+|1α2|+2uo2/dio2\underline{C}^{-1}(O_{i,j}(o,u_{o}))|_{\theta_{ji}=\pm\frac{\pi}{2}}=\sqrt{\frac{1+\alpha^{2}-|1-\alpha^{2}|}{1+\alpha^{2}+|1-\alpha^{2}|+2u_{o}^{2}/d_{io}^{2}}}. Then, when θji=±π2\theta_{ji}=\pm\frac{\pi}{2},

C¯1(Oi,j(o,uo))α|θji=±π2=0\frac{\partial\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\alpha}|_{\theta_{ji}=\pm\frac{\pi}{2}=0}
α(1α2)(1+α2+|1α2|+2uo2/dio2)2.5=0,\Rightarrow\frac{\alpha(1-\alpha^{2})}{(1+\alpha^{2}+|1-\alpha^{2}|+2u_{o}^{2}/d_{io}^{2})^{2.5}}=0,
α=0,1,α.\Rightarrow\alpha=0,1,\alpha\to\infty.

and

2C¯1(Oi,j(o,uo))α2|θji=±π2,α=0,α>0.\frac{\partial^{2}\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\alpha^{2}}|_{\theta_{ji}=\pm\frac{\pi}{2},\alpha=0,\alpha\to\infty}>0.
2C¯1(Oi,j(o,uo))α2|θji=±π2,α=1<0.\frac{\partial^{2}\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\alpha^{2}}|_{\theta_{ji}=\pm\frac{\pi}{2},\alpha=1}<0.

Therefore, C¯1(Oi,j(o,uo))|θji=±π2\underline{C}^{-1}(O_{i,j}(o,u_{o}))|_{\theta_{ji}=\pm\frac{\pi}{2}} reaches its minimum 0, at α=0\alpha=0, or α\alpha\to\infty, and reaches its maximum 11+uo2/dio2\sqrt{\frac{1}{1+u_{o}^{2}/d_{io}^{2}}}, at α=1\alpha=1, as shown in Figure 2 where, C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) reaches its maximum at α=1\alpha=1 and θji=±π2{\theta_{ji}=\pm\frac{\pi}{2}}.

Note that, when θji=0,π\theta_{ji}=0,\pi, C¯1(Oi,j(o,uo))\underline{C}^{-1}(O_{i,j}(o,u_{o})) is always null and not influenced by α\alpha. Besides, for a fixed θji0,±π2,π\theta_{ji}\neq 0,\pm\frac{\pi}{2},\pi, the minimum extreme point w.r.t. α\alpha still happens when α=0\alpha=0 or α\alpha\to\infty, but the maximum extreme point is not at α=1\alpha=1, which is shown is the following derivations.

C¯1(Oi,j(o,uo))α|θji0,±π2,π=0\frac{\partial\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\alpha}|_{\theta_{ji}\neq 0,\pm\frac{\pi}{2},\pi}=0
α[(1α2)(1cos(2θji))(1+α2+1+α4+2α2cos(2θji)+2uo2/dio2)2.5\displaystyle\Rightarrow\frac{\alpha[(1-\alpha^{2})(1-\cos(2\theta_{ji}))}{(1+\alpha^{2}+\sqrt{1+\alpha^{4}+2\alpha^{2}\cos(2\theta_{ji})}+2u_{o}^{2}/d_{io}^{2})^{2.5}}
+uo2/dio2(1+α4+2α2cos(2θji)(α2+cos(2θji)))](1+α2+1+α4+2α2cos(2θji)+2uo2/dio2)2.5\displaystyle\frac{+u_{o}^{2}/d_{io}^{2}(\sqrt{1+\alpha^{4}+2\alpha^{2}\cos(2\theta_{ji})}-(\alpha^{2}+\cos(2\theta_{ji})))]}{(1+\alpha^{2}+\sqrt{1+\alpha^{4}+2\alpha^{2}\cos(2\theta_{ji})}+2u_{o}^{2}/d_{io}^{2})^{2.5}}
=0.\displaystyle=0.
α=0,α.\displaystyle\Rightarrow\alpha=0,\alpha\to\infty.

and \Rightarrow

(1α2)(1cos(2θji))+\displaystyle(1-\alpha^{2})(1-\cos(2\theta_{ji}))+
uo2dio2(1+α4+2α2cos(2θji)(α2+cos(2θji)))=0.\displaystyle\frac{u_{o}^{2}}{d_{io}^{2}}(\sqrt{1+\alpha^{4}+2\alpha^{2}\cos(2\theta_{ji})}-(\alpha^{2}+\cos(2\theta_{ji})))=0. (39)

Since

2C¯1(Oi,j(o,uo))α2|θji0,±π2,π,α=0,α>0,\frac{\partial^{2}\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\alpha^{2}}|_{\theta_{ji}\neq 0,\pm\frac{\pi}{2},\pi,\alpha=0,\alpha\to\infty}>0,

C¯1(Oi,j(o,uo))α|θji0,±π2,π\frac{\partial\underline{C}^{-1}(O_{i,j}(o,u_{o}))}{\partial\alpha}|_{\theta_{ji}\neq 0,\pm\frac{\pi}{2},\pi} also reaches its minimum at α=0\alpha=0 or α\alpha\to\infty. However, α=1\alpha=1 cannot make Equation 39 established when θji0,±π2,π\theta_{ji}\neq 0,\pm\frac{\pi}{2},\pi, and thus, the extreme point is not at α=1\alpha=1.