This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Mismatched Estimation
in the Distance Geometry Problem

Mahmoud Abdelkhalek, Dror Baron, and Chau-Wai Wong Electrical and Computer Engineering, North Carolina State University
Email: {maabdelk, barondror, chauwai.wong}@ncsu.edu
Abstract

We investigate mismatched estimation in the context of the distance geometry problem (DGP). In the DGP, for a set of points, we are given noisy measurements of pairwise distances between the points, and our objective is to determine the geometric locations of the points. A common approach to deal with noisy measurements of pairwise distances is to compute least-squares estimates of the locations of the points. However, these least-squares estimates are likely to be suboptimal, because they do not necessarily maximize the correct likelihood function. In this paper, we argue that more accurate estimates can be obtained when an estimation procedure using the correct likelihood function of noisy measurements is performed. Our numerical results demonstrate that least-squares estimates can be suboptimal by several dB.

I Introduction

The distance geometry problem (DGP) [1, 2, 3] aims to determine the locations of points in Euclidean space given measurements of the distances between these points. The DGP arises in many important applications. In wireless networks with moving devices in two-dimensional (2D) planes, the positions of these devices can be estimated by solving the DGPΒ [4]. In computational biology, solving the DGP can allow us to determine the locations of atoms in molecules [5]. In robotics, a simple robotic movement can be modeled as a pair of rigid bars rotating along a joint. Whether these bars can flex along a joint to reach a specific point in 3D space can be determined by solving the DGP [6].

The main challenge for the DGP is that pairwise distance measurements are noisy [1]. Because of the noise, it is common to find a set of locations of points that minimizes the sum of squared errors (SSE) between the pairwise distances of a target structure and noisy measurements of these pairwise distances [7]. When the noisy measurements of pairwise distances are Gaussian distributed, minimizing the SSE function yields a maximum likelihood (ML) estimate of a structure of points. However, in practice, it is unlikely that the measurements are exactly Gaussian distributed, and so the SSE function is unlikely to yield an ML estimate. In this work, we will provide better estimation quality by performing ML estimation using the correct noise distribution. We call our approach matched DGP estimation. In contrast, much of the prior art performed mismatched DGP estimation (details below).

Problem formulation. Let 𝒳={𝐱1,𝐱2,…,𝐱N}\mathcal{X}=\{\mathbf{x}_{1},\mathbf{x}_{2},\dots,\mathbf{x}_{N}\} be an unobserved set of locations of NN points in KK-dimensional Euclidean space ℝK\mathbb{R}^{K}. Additionally, let ℰ​(𝒳)={(i,j)∣(i,j)∈{1,…,N}2,i<j}\mathcal{E}(\mathcal{X})=\{(i,j)\mid(i,j)\in\{1,\dots,N\}^{2},i<j\} be the set of edges connecting every pair of points in 𝒳\mathcal{X}. Note that |ℰ​(𝒳)|=12​N​(Nβˆ’1)|\mathcal{E}(\mathcal{X})|=\frac{1}{2}N(N-1), where |β‹…||\cdot| denotes the cardinality of a set. We define the length di​j𝒳d_{ij}^{\mathcal{X}} of an edge (i,j)βˆˆβ„°β€‹(𝒳)(i,j)\in\mathcal{E}(\mathcal{X}) as the β„“2\ell_{2} norm of 𝐱iβˆ’π±j\mathbf{x}_{i}-\mathbf{x}_{j}, di​j𝒳=‖𝐱iβˆ’π±jβ€–2d_{ij}^{\mathcal{X}}=\left\|\mathbf{x}_{i}-\mathbf{x}_{j}\right\|_{2}. We define Mi​jβˆˆβ„•M_{ij}\in\mathbb{N}, where β„•\mathbb{N} is the set of nonnegative integers, as the number of noisy measurements of the length of edge (i,j)βˆˆβ„°β€‹(𝒳)(i,j)\in\mathcal{E}(\mathcal{X}). An example structure 𝒳\mathcal{X} is shown in Fig.Β 1. In the DGP, our objective is to determine 𝒳\mathcal{X} given noisy measurements of lengths of edges in ℰ​(𝒳)\mathcal{E}(\mathcal{X}).

Estimating the locations of the points. We estimate 𝒳\mathcal{X} using the method of ML. Let the random variable Yi​j[m]βˆˆβ„+Y_{ij}^{[m]}\in\mathbb{R}^{+} represent the mmth noisy measurement of the length of edge (i,j)(i,j), where a realization of Yi​j[m]Y_{ij}^{[m]} is denoted by yi​j[m]y_{ij}^{[m]}. For example, y1,2[8]y_{1,2}^{[8]} is the 88th measurement of the length of edge (1,2)(1,2). We assume that Mi​j=MM_{ij}=M for every (i,j)βˆˆβ„°β€‹(𝒳)(i,j)\in\mathcal{E}(\mathcal{X}). Let the set of all Yi​j[m]Y_{ij}^{[m]} be, 𝒴=⋃(i,j)βˆˆβ„°β€‹(𝒳)𝒴i​j\mathscr{Y}=\bigcup_{(i,j)\in\mathcal{E}(\mathcal{X})}\mathscr{Y}_{ij}, where 𝒴i​j={Yi​j[m]}m=1M\mathscr{Y}_{ij}=\left\{Y_{ij}^{[m]}\right\}_{m=1}^{M}. The sets π“Ž\mathscr{y} and π“Ži​j\mathscr{y}_{ij} are defined similarly for yi​j[m]y_{ij}^{[m]}. Our objective is to derive an expression for the likelihood function f𝒴​(π“Žβˆ£dℰ​(𝒳))f_{\mathscr{Y}}(\mathscr{y}\mid d_{\mathcal{E}(\mathcal{X})}), where dℰ​(𝒳)={di​j𝒳}(i,j)βˆˆβ„°β€‹(𝒳)d_{\mathcal{E}(\mathcal{X})}=\left\{d_{ij}^{\mathcal{X}}\right\}_{(i,j)\in\mathcal{E}(\mathcal{X})}.

Suppose that the measurements for edge (i1,j1)(i_{1},j_{1}) are independent of the measurements for edge (i2,j2)(i_{2},j_{2}) for every (i1,j1),(i2,j2)βˆˆβ„°β€‹(𝒳)(i_{1},j_{1}),(i_{2},j_{2})\in\mathcal{E}(\mathcal{X}), and that 𝔼​[Yi​j[m]]=di​j𝒳\mathbb{E}\left[Y_{ij}^{[m]}\right]=d_{ij}^{\mathcal{X}} for every m=1,…,Mm=1,\dots,M. Under these assumptions, the likelihood can be expressed as, f𝒴(π“Žβˆ£dℰ​(𝒳))=∏(i,j)βˆˆβ„°β€‹(𝒳)f𝒴i​j(π“Ži​j|di​j𝒳)f_{\mathscr{Y}}(\mathscr{y}\mid d_{\mathcal{E}(\mathcal{X})})=\prod_{(i,j)\in\mathcal{E}(\mathcal{X})}f_{\mathscr{Y}_{ij}}\left(\mathscr{y}_{ij}\bigm{|}d_{ij}^{\mathcal{X}}\right). Additionally, suppose that all measurements associated with edge (i,j)(i,j) are independent and identically distributed (i.i.d.). The likelihood can be further simplified,

f𝒴(π“Žβˆ£dℰ​(𝒳))=∏(i,j)βˆˆβ„°β€‹(𝒳)∏m=1MfYi​j[m](yi​j[m]|di​j𝒳),f_{\mathscr{Y}}\left(\mathscr{y}\mid d_{\mathcal{E}(\mathcal{X})}\right)=\prod_{(i,j)\in\mathcal{E}(\mathcal{X})}\prod_{m=1}^{M}f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}}\right), (1)

where fYi​j[m](yi​j[m]|di​j𝒳)f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}}\right) is the true marginal pdf of Yi​j[m]Y_{ij}^{[m]}. In the rest of the paper, we will use fYi​j[m](yi​j[m]|di​j𝒳,𝜽i​j)f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}},\bm{\theta}_{ij}\right), where 𝜽i​j\bm{\theta}_{ij} is a vector of shape parameters, to denote the marginal pdf for Yi​j[m]Y_{ij}^{[m]}. When a single shape parameter describes this marginal pdf, the scalar version ΞΈi​j\theta_{ij} will be used instead.

Given a set of noisy edge length measurements π“Ž\mathscr{y}, an ML estimate of 𝒳\mathcal{X} can be obtained by maximizing (1). This case will be referred to as matched DGP estimation. In practice, we often do not know what parametric form the marginal pdf of Yi​j[m]Y_{ij}^{[m]} takes. In this case, we would modifyΒ  ​(1) by replacing the true marginal pdf of Yi​j[m]Y_{ij}^{[m]} with an assumed version, fYi​j[m]βˆ—(yi​j[m]|di​j𝒳)f_{Y_{ij}^{[m]}}^{*}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}}\right). If fYi​j[m]βˆ—(yi​j[m]|di​j𝒳)β‰ fYi​j[m](yi​j[m]|di​j𝒳)f_{Y_{ij}^{[m]}}^{*}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}}\right)\neq f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}}\right), then we refer to the modified version of the likelihood as the mismatched likelihood. Moreover, an estimate of 𝒳\mathcal{X} that is obtained by maximizing the mismatched likelihood will not necessarily be an ML estimate. We refer to this case as mismatched DGP estimation. We are interested in the difference in quality of estimates of 𝒳\mathcal{X} in the matched and mismatched cases.

Refer to caption
Figure 1: Example structure 𝒳\mathcal{X} comprised of 3 points, 𝐱1,𝐱2,𝐱3βˆˆβ„2\mathbf{x}_{1},\mathbf{x}_{2},\mathbf{x}_{3}\in\mathbb{R}^{2}.

Limitations of least-squares. Suppose that Yi​j[m]Y_{ij}^{[m]} follows a Gaussian distribution with mean di​j𝒳d_{ij}^{\mathcal{X}} and variance ΞΈi​j\theta_{ij}, namely,

fYi​j[m](yi​j[m]|di​j𝒳,ΞΈi​j)=12​π​θi​jexp(βˆ’(yi​j[m]βˆ’di​j𝒳)22​θi​j).f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}},\theta_{ij}\right)=\frac{1}{\sqrt{2\pi\theta_{ij}}}\exp\left(-\frac{(y_{ij}^{[m]}-d_{ij}^{\mathcal{X}})^{2}}{2\theta_{ij}}\right).

Then, maximizing the likelihood f𝒴​(π“Žβˆ£dℰ​(𝒳))f_{\mathscr{Y}}\left(\mathscr{y}\mid d_{\mathcal{E}(\mathcal{X})}\right) with respect to (w.r.t.) 𝒳\mathcal{X} is equivalent to minimizing the SSE w.r.t. 𝒳\mathcal{X}. Several approaches to solve the DGP in the case of noisy measurements use variants of the SSE function [3, 8, 9]. A survey of prior art appears in Sec.Β II.

The aforementioned equivalence between minimizing the SSE function and maximizing the likelihood derived in (1) is only true when Yi​j[m]Y_{ij}^{[m]} is Gaussian. In practice, depending on the measurement technique, it is unlikely that the noisy measurement Yi​j[m]Y_{ij}^{[m]} is Gaussian distributed. Therefore, it is unlikely that minimizing the SSE function will yield an ML estimate of a structure.

Contributions. In this paper, we empirically show that, compared to mismatched DGP estimation, matched DGP estimation can offer an improvement of several dB. Moreover, our contribution is that, to the best of our knowledge, and in the context of the DGP, we are the first to focus on ML estimation beyond Gaussian-distributed measurements. The rest of the paper is structured as follows: we review the prior art in Sec.Β II. The different marginal pdfs of Yi​j[m]Y_{ij}^{[m]} that will be considered, and how estimates of 𝒳\mathcal{X} will be evaluated, are discussed in Sec.Β III. Experimental results are presented in Sec.Β IV. Finally, we summarize our results and describe future work in Sec.Β V​.

II Related Work

One of the earliest discussions of cost functions related to the solution of the DGP was due to Crippen and Havel [3]. In their work, they used a variant of the SSE function, which we refer to as SSE-CH, to estimate molecular structures with physical constraints. Although others have empirically shown that SSE-CH yields acceptable results [3], the resulting estimate of a structure need not be an ML estimate.

Subsequent papers implement modifications to the SSE-CH cost function. For example, MorΓ© and Wu [9] adopted a continuation approach: the SSE-CH cost function is repeatedly smoothed via a convolution with a Gaussian kernel. After each successive smoothing stage, a minimizer of the smoothed SSE-CH is obtained. Eventually, the minimizers are traced back to the original function [9, p. 827]. Although this approach yielded promising results, the underlying cost function that is smoothed is still the SSE function. Moreover, there is no guarantee that the resulting estimate of 𝒳\mathcal{X} is an ML estimate.

Souza et al. [10] attempt to make SSE-CH differentiable by approximating certain parts using hyperbolic functions. Souza et al. [11] made further improvements by applying these approximations to the geometric buildup algorithm proposed by Luo and Wu [12]. However, the underlying cost function is still a variant of the SSE function, and so there is no guarantee of obtaining ML estimates. For a detailed review of approaches to solve the DGP, see [7]. More recently, Gopalkrishnan et al. [13] optimize a variation of SSE-CH to reconstruct DNA molecules from pairwise distance measurements. Giamou et al. [14] also optimize a variation of SSE-CH to determine the joint angles of a robot given the robot’s forward kinematics and desired poses. However, because variations of SSE-CH are used, it is unlikely that estimates obtained in these applications are ML estimates.

III Noise Distributions and Evaluation

We will consider different forms of the marginal pdf of Yi​j[m]Y_{ij}^{[m]}, as included in (1), that result in a likelihood f𝒴​(π“Žβˆ£dℰ​(𝒳))f_{\mathscr{Y}}\left(\mathscr{y}\mid d_{\mathcal{E}(\mathcal{X})}\right) that, when maximized w.r.t. 𝒳\mathcal{X}, yields ML estimates. It was previously shown that if Yi​j[m]Y_{ij}^{[m]} follows a Gaussian distribution, then minimizing the SSE function is equivalent to maximizing the likelihood in (1). We consider the cases when Yi​j[m]Y_{ij}^{[m]} follows either a Laplace or non-standardized Student’s tt (NSST) distribution.

If Yi​j[m]Y_{ij}^{[m]} follows a Laplace distribution with mean di​j𝒳d_{ij}^{\mathcal{X}} and variance 2​θi​j22\theta_{ij}^{2}, then

fYi​j[m](yi​j[m]|di​j𝒳,ΞΈi​j)=12​θi​jeβˆ’|yi​j[m]βˆ’di​j𝒳|/ΞΈi​j.f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}},\theta_{ij}\right)=\frac{1}{2\theta_{ij}}e^{-\left\lvert y_{ij}^{[m]}-d_{ij}^{\mathcal{X}}\right\rvert\big{/}\theta_{ij}}. (2)

If Yi​j[m]Y_{ij}^{[m]} follows an NSST distribution with mean di​j𝒳d_{ij}^{\mathcal{X}} and variance bi​j2​νi​j/(Ξ½i​jβˆ’2)b_{ij}^{2}\nu_{ij}/(\nu_{ij}-2), assuming Ξ½i​j>2\nu_{ij}>2, then an expression for the marginal pdf of Yi​j[m]Y_{ij}^{[m]} can be derived as follows. Let WW follow a Student’s tt distribution with parameter Ξ½i​j\nu_{ij}, which is a positive integer. The pdf of WW is

fW​(w∣νi​j)=Γ​((Ξ½i​j+1)/2)Γ​(Ξ½i​j/2)​νi​j​π​(1+(w2/Ξ½i​j))βˆ’(Ξ½i​j+1)/2,f_{W}(w\mid\nu_{ij})=\frac{\Gamma\left((\nu_{ij}+1)/2\right)}{\Gamma\left(\nu_{ij}/2\right)\sqrt{\nu_{ij}\pi}}\left(1+(w^{2}/\nu_{ij})\right)^{-(\nu_{ij}+1)/2},

where Γ​(z)=∫0∞tzβˆ’1​eβˆ’t​𝑑t\Gamma(z)=\int_{0}^{\infty}t^{z-1}e^{-t}\,dt for every 0<zβˆˆβ„0<z\in\mathbb{R}. If Yi​j[m]=bi​j​W+di​j𝒳Y_{ij}^{[m]}=b_{ij}W+d_{ij}^{\mathcal{X}}, then

fYi​j[m](yi​j[m]|di​j𝒳,𝜽i​j)=1bi​jfW(yi​j[m]βˆ’di​j𝒳bi​j|Ξ½i​j),f_{Y_{ij}^{[m]}}\left(y_{ij}^{[m]}\Bigm{|}d_{ij}^{\mathcal{X}},\bm{\theta}_{ij}\right)=\frac{1}{b_{ij}}f_{W}\left(\frac{y_{ij}^{[m]}-d_{ij}^{\mathcal{X}}}{b_{ij}}\Biggm{|}\nu_{ij}\right), (3)

where 𝜽i​j=(Ξ½i​j,bi​j)\bm{\theta}_{ij}=(\nu_{ij},b_{ij}).

Evaluation. Given an estimate 𝒳^\widehat{\mathcal{X}} of 𝒳\mathcal{X}, if we rotate, reflect, and/or translate the points in 𝒳^\widehat{\mathcal{X}} by an equal amount, then 𝒳^\widehat{\mathcal{X}} will still have the same edge lengths as 𝒳\mathcal{X}. Therefore, a unique solution to the DGP cannot exist without additional constraints. Moreover, the proximity of 𝒳^\widehat{\mathcal{X}} to 𝒳\mathcal{X} will need to be evaluated up to rotation, translation, and reflection. Let 𝐗\mathbf{X} and 𝐗^\widehat{\mathbf{X}} be KΓ—NK\times N matrices, where the nnth columns of 𝐗\mathbf{X} and 𝐗^\widehat{\mathbf{X}} correspond to the nnth points in 𝒳\mathcal{X} and 𝒳^\widehat{\mathcal{X}}, respectively. We first translate the column vectors in 𝐗\mathbf{X} and 𝐗^\widehat{\mathbf{X}} so that they are centered at the origin by computing 𝐗c=π—βˆ’π—πŸ/N\mathbf{X}_{c}=\mathbf{X}-\mathbf{X}\mathbf{1}\big{/}N and 𝐗^c=𝐗^βˆ’π—^β€‹πŸ/N\widehat{\mathbf{X}}_{c}=\widehat{\mathbf{X}}-\widehat{\mathbf{X}}\mathbf{1}\big{/}N, where 𝟏\mathbf{1} is an NΓ—NN\times N matrix of ones.

Next, how close 𝒳^\widehat{\mathcal{X}} is to 𝒳\mathcal{X} is equivalent to the solution of the orthogonal Procrustes problem (OPP) [15],

min𝐑\displaystyle\min_{\mathbf{R}} ‖𝐑​𝐗^cβˆ’π—cβ€–F,\displaystyle\left\|\mathbf{R}\widehat{\mathbf{X}}_{c}-\mathbf{X}_{c}\right\|_{F}, (4)
s.t. 𝐑T​𝐑=𝐈,\displaystyle\mathbf{R}^{T}\mathbf{R}=\mathbf{I},

where βˆ₯β‹…βˆ₯F\left\|\cdot\right\|_{F} is the Frobenius norm, which for a KΓ—NK\times N matrix, 𝐀=(ai​j)\mathbf{A}=(a_{ij}), is defined as, ‖𝐀‖F=(βˆ‘k=1Kβˆ‘n=1N|ai​j|2)1/2\left\|\mathbf{A}\right\|_{F}=\left(\sum_{k=1}^{K}\sum_{n=1}^{N}|a_{ij}|^{2}\right)^{1/2}. The objective of the OPP is to determine the orthonormal matrix 𝐑\mathbf{R} that most closely maps 𝐗^c\widehat{\mathbf{X}}_{c} to 𝐗c\mathbf{X}_{c}. We will refer to the value obtained by solving the OPP for an estimate 𝒳^\widehat{\mathcal{X}}, divided by the number of points NN, as the OPP loss associated with 𝒳^\widehat{\mathcal{X}}. As our objective is to show that an ML estimate 𝒳^ML\widehat{\mathcal{X}}_{\text{ML}} of 𝒳\mathcal{X} more closely approximates 𝒳\mathcal{X} than a non-ML estimate 𝒳^\widehat{\mathcal{X}}, we will compare the OPP losses of 𝒳^ML\widehat{\mathcal{X}}_{\text{ML}} and 𝒳^\widehat{\mathcal{X}}.

IV Numerical Results

Refer to caption
((a))
Refer to caption
((b))
Figure 2: We illustrate 2 of the 8 triangles that were considered. Ground truth structures are labeled 𝒳\mathcal{X}, while example estimates, together with their associated OPP losses, are also shown. We observe that estimates with lower OPP losses more closely approximate the structure 𝒳\mathcal{X}.

We begin our discussion with numerical results on triangles, which are structures consisting of 3 points (see Fig.Β 2). Later, we demonstrate analogous numerical results for structures with 10 points (see Fig.Β 5). In our experiments, we let Yi​j[m]Y_{ij}^{[m]} follow either a Laplace (2) or NSST (3) distribution. If Yi​j[m]Y_{ij}^{[m]} follows a Laplace (respectively, NSST) distribution, then the marginal pdf of Yi​j[m]Y_{ij}^{[m]} in the likelihood (1) is varied between the Laplace (respectively, NSST) pdf in the matched case and the Gaussian pdf in the mismatched case.

The variance of the noisy measurements was changed as a function of different signal-to-noise ratio (SNR) values. If the distribution from which the i.i.d. noisy measurements associated with edge (i,j)(i,j) were obtained has variance Οƒi​j2\sigma_{ij}^{2}, and if we let σ𝒳\sigma_{\mathcal{X}} be the average edge length of a structure 𝒳\mathcal{X}, then SNR=10​log10⁑(σ𝒳2/Οƒi​j2)\text{SNR}=10\log_{10}\left(\sigma_{\mathcal{X}}^{2}/\sigma_{ij}^{2}\right), where SNR is quantified in decibel (dB) units. We leave the case where Οƒi​j2\sigma_{ij}^{2} depends on edge (i,j)(i,j) for future work. The SNR was varied from –20 dB to 20 dB in steps of 5 dB. For each SNR value, we sample a set of noisy measurements, maximize the likelihood111An implementation of the L-BFGS algorithm [16] that is available in PyTorch [17] was used to maximize the likelihood. in (1) for the Laplace (or NSST) cases, and then report the OPP loss associated with an estimated structure 𝒳^\widehat{\mathcal{X}} for each case.

Refer to caption
((a))
Refer to caption
((b))
Refer to caption
((c))
Figure 3: Distributions of OPP losses for the 8 triangles when Yi​j[m]Y_{ij}^{[m]} follows a LABEL:sub@fig:3pts_M10_laplace Laplace or LABEL:sub@fig:3pts_M10_nsst NSST distribution, and when matched and mismatched (Gaussian) likelihood (LL) functions are used. Each box represents the percentiles (bottom to top): 10, 25, 50, 75, and 90. Distributions of pairwise differences in OPP loss when Yi​j[m]Y_{ij}^{[m]} follows a Laplace or NSST distribution are shown in LABEL:sub@fig:3pts_M10_diff. Pairwise differences are computed by subtracting the OPP loss for the matched estimate from the corresponding OPP loss for the mismatched (Gaussian LL function) estimate.

Triangles. We first considered 8 triangles in 2D. Two of these triangles, together with example estimates and their associated OPP losses, are shown in Fig.Β 2. For each SNR value, and for each triangle, we sampled M=10M=10 noisy measurements per edge 100 times. For each sample, we computed an estimate 𝒳^\widehat{\mathcal{X}} using different likelihoods and then computed the OPP loss associated with 𝒳^\widehat{\mathcal{X}}. The distributions of these OPP losses are shown in Fig.Β 3(a) and Fig.Β 3(b). We also computed the pairwise difference in OPP losses resulting from mismatched and matched estimation. The distributions of these differences are shown in Fig.Β 3(c).

We observe that, in general, the distribution of OPP losses associated with the likelihood that is matched to the noise distribution is shifted lower than the distribution associated with the likelihood that is not matched to the noise distribution. Moreover, in several instances, the increase in median OPP loss due to a 3–4 dB decrease in SNR can approximately be compensated for by choosing to use the likelihood that is matched to the noise distribution.

Refer to caption
((a))
Refer to caption
((b))
Figure 4: Median OPP losses for the 8 triangles and for different MM values when Yi​j[m]Y_{ij}^{[m]} follows a LABEL:sub@fig:3pts_Ms_laplace Laplace or LABEL:sub@fig:3pts_Ms_nsst NSST distribution, and when matched or mismatched (Gaussian) likelihood functions are used.

We then repeated this experiment for different values of MM, the number of noisy measurements per edge. The median OPP losses for M=10,25,50,100M=10,25,50,100 are shown in Fig.Β 4. We observe that, in general, maximizing the likelihood that is matched to the noise distribution offers an improvement over mismatched DGP estimation comparable to the improvement obtained by doubling the number of noisy measurements per edge.

Refer to caption
Figure 5: One of the 30 10-point structures that were considered. Example estimates are omitted for clarity.
Refer to caption
((a))
Refer to caption
((b))
Refer to caption
((c))
Figure 6: Distributions of OPP losses for the 30 10-point structures when Yi​j[m]Y_{ij}^{[m]} follows a LABEL:sub@fig:10pts_M10_laplace Laplace or LABEL:sub@fig:10pts_M10_nsst NSST distribution, and when matched and mismatched (Gaussian) likelihood (LL) functions are used. Percentiles are marked as in Fig.Β 3. Distributions of pairwise differences in OPP loss when Yi​j[m]Y_{ij}^{[m]} follows a Laplace or NSST distribution are shown in LABEL:sub@fig:10pts_M10_diff. Pairwise differences are computed by subtracting the OPP loss for the matched estimate from the corresponding OPP loss for the mismatched (Gaussian LL function) estimate.

10 points. To further support our results, we repeated our experiments for 30 10-point structures in 2D. One of these structures is shown in Fig.Β 5. The distributions of OPP losses are shown in Fig.Β 6(a) and Fig.Β 6(b), and the distributions of pairwise differences in OPP losses for the mismatched and matched cases are shown in Fig.Β 6(c). Similar to the case of triangles, we observe that, in general, the distribution of OPP losses associated with the likelihood that is matched to the noise distribution is lower than the distribution associated with the likelihood that is not matched to the noise distribution.

V Conclusions and Future Work

In this paper, we have explored the impact of matching the likelihood function to the noise distribution on estimation accuracy in the context of the distance geometry problem (DGP). We have empirically found that matching the likelihood function to the noise distribution can offer a 3–4 dB compensation for the loss in estimation accuracy due to noisier measurements and smaller sample sizes. Moreover, matched estimation can be helpful in these scenarios, which often arise in practice.

There are several directions for future work that can be explored. For example, in practice, measurements may not be available for some edges, so the case where the numbers of edge measurements are unequal may also be considered. We may also explore the application of our approach to real-world measurements. Finally, a theoretical evaluation of the mismatch between the likelihood function and the noise distribution would be an insightful endeavor.

Acknowledgments

We thank Vaibhav Choudhary and Nikhil Gopalkrishnan for informative discussions.

References

  • [1] Leo Liberti and Carlile Lavor β€œThe Distance Geometry Problem” In Euclidean Distance Geometry: An Introduction Cham: Springer International Publishing, 2017, pp. 9–18 DOI: 10.1007/978-3-319-60792-4Λ™2
  • [2] Zachary Voller and Zhijun Wu β€œDistance Geometry Methods for Protein Structure Determination” In Distance Geometry: Theory, Methods, and Applications New York, NY: Springer, 2013, pp. 139–159 DOI: 10.1007/978-1-4614-5128-0Λ™8
  • [3] G.. Crippen and Timothy F. Havel β€œDistance Geometry and Molecular Conformation” Research Studies Press, 1988 GOOGLEBOOKS: 5CPwAAAAMAAJ
  • [4] T. Eren et al. β€œRigidity, Computation, and Randomization in Network Localization” In IEEE INFOCOM 4, 2004, pp. 2673–2684 DOI: 10.1109/INFCOM.2004.1354686
  • [5] T.. Havel β€œAn Evaluation of Computational Strategies for Use in the Determination of Protein Structure from Distance Constraints Obtained by Nuclear Magnetic Resonance” In Prog. Biophys. Mol. Biol. 56.1, 1991, pp. 43–78 DOI: 10.1016/0079-6107(91)90007-F
  • [6] Leo Liberti and Carlile Lavor β€œOpen Research Areas in Distance Geometry” In Open Problems in Optimization and Data Analysis Springer, 2018, pp. 183–223
  • [7] Carlile Lavor, Leo Liberti and Nelson Maculan β€œAn Overview of Distinct Approaches for the Molecular Distance Geometry Problem” In Encyclopedia of Optimization Boston, MA: Springer US, 2009, pp. 2304–2311 DOI: 10.1007/978-0-387-74759-0Λ™400
  • [8] Bruce Hendrickson β€œThe Molecule Problem: Exploiting Structure in Global Optimization” In SIAM J. Optim. 5.4 Society for Industrial and Applied Mathematics, 1995, pp. 835–857 DOI: 10.1137/0805040
  • [9] Jorge J. MorΓ© and Zhijun Wu β€œGlobal Continuation for Distance Geometry Problems” In SIAM J. Optim. 7.3 Society for Industrial and Applied Mathematics, 1997, pp. 814–836 DOI: 10.1137/S1052623495283024
  • [10] Michael Souza, Adilson Elias Xavier, Carlile Lavor and Nelson Maculan β€œHyperbolic Smoothing and Penalty Techniques Applied to Molecular Structure Determination” In Operations Research Letters 39.6, 2011, pp. 461–465 DOI: 10.1016/j.orl.2011.07.007
  • [11] Michael Souza, Carlile Lavor, Albert Muritiba and Nelson Maculan β€œSolving the Molecular Distance Geometry Problem with Inaccurate Distance Data” In BMC Bioinform. 14, 2013, pp. S7 DOI: 10.1186/1471-2105-14-S9-S7
  • [12] Xin-Long Luo and Zhi-Jun Wu β€œLeast-Squares Approximations in Geometric Buildup for Solving Distance Geometry Problems” In J. Optim. Theory Appl. 149.3, 2011, pp. 580–598 DOI: 10.1007/s10957-011-9806-6
  • [13] Nikhil Gopalkrishnan, Sukanya Punthambaker, Thomas E Schaus, George M Church and Peng Yin β€œA DNA nanoscope that identifies and precisely localizes over a hundred unique molecular features with nanometer accuracy” In bioRxiv, 2020
  • [14] Matthew Giamou et al. β€œConvex Iteration for Distance-Geometric Inverse Kinematics” In IEEE Robot. Autom. Lett. 7.2, 2022, pp. 1952–1959 DOI: 10.1109/LRA.2022.3141763
  • [15] Peter H. SchΓΆnemann β€œA Generalized Solution of the Orthogonal Procrustes Problem” In Psychometrika 31.1, 1996, pp. 1–10 DOI: 10.1007/BF02289451
  • [16] β€œQuasi-Newton Methods” In Numerical Optimization, Springer Series in Operations Research and Financial Engineering New York, NY: Springer, 2006, pp. 135–163 DOI: 10.1007/978-0-387-40065-5Λ™6
  • [17] Adam Paszke et al. β€œPytorch: An imperative style, high-performance deep learning library” In Advances in Neural Information Processing Systems 32, 2019