This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

SiML: Sieved Maximum Likelihood for Array Signal Processing

Abstract

Stochastic Maximum Likelihood (SML) is a popular direction of arrival (DOA) estimation technique in array signal processing. It is a parametric method that jointly estimates signal and instrument noise by maximum likelihood, achieving excellent statistical performance. Some drawbacks are the computational overhead as well as the limitation to a point-source data model with fewer sources than sensors. In this work, we propose a Sieved Maximum Likelihood (SiML) method. It uses a general functional data model, allowing an unrestricted number of arbitrarily-shaped sources to be recovered. To this end, we leverage functional analysis tools and express the data in terms of an infinite-dimensional sampling operator acting on a Gaussian random function. We show that SiML is computationally more efficient than traditional SML, resilient to noise, and results in much better accuracy than spectral-based methods.

©Copyright 2021 IEEE. Published in ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), scheduled for 6-11 June 2021 in Toronto, Ontario, Canada. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works, must be obtained from the IEEE. Contact: Manager, Copyrights and Permissions / IEEE Service Center / 445 Hoes Lane / P.O. Box 1331 / Piscataway, NJ 08855-1331, USA. Telephone: + Intl. 908-562-3966.

Index Terms— stochastic maximum likelihood, sieved maximum likelihood, spatially extended sources, random fields, sampling operator, array signal processing.

1 Introduction

Array signal processing [1, 2, 3] is primarily concerned with the sensing, processing and estimation of random wavefields (electromagnetic or mechanic). Techniques from array signal processing are used in myriad of applications, including for example acoustics [4, 5], radio-interferometry [6, 7], radar and sonar systems [2, 8], wireless networks [9, 10, 11], and medical imagery [12, 13]. The sensing devices in all those applications consist of large networks of sensors, called sensor arrays or phased-arrays.

A common task in array signal processing consists of estimating the intensity field of an emitting wavefield. The various algorithms available in the literature for this purpose divide into two categories [1]: spectral-based and parametric methods. Spectral-based ones estimate the intensity field by “steering” the array towards particular directions in space and evaluating the output power. The intensity field is thus recovered sequentially by scanning via beamforming [1, 8, 11] a grid covering the field of view. Famous beamformers include Bartlett, also known as Matched Beamforming (MB), and Capon, which is often called Minimum Variance Distortionless Response (MVDR) [1]. These are extremely simple to implement, computationally attractive, and quite generic with little structural or distributional assumptions on the sensed wavefield. They are however limited in terms of accuracy, in particular for extreme acquisition conditions such as small sample size, low signal-to-noise ratio and/or spatial correlation in the wavefield [1].

Parametric methods, on the other hand, attempt to overcome those limitations using a statistical model for the instrument noise and the unknown wavefield. Typically, the thermal noise at each sensor is modelled as an additive Gaussian white random process while the wavefield is assumed to be the incoherent sum of QQ point sources with amplitudes distributed according to a multivariate complex Gaussian distribution, where QQ is strictly smaller than the total number of sensors composing the array. In this context the problem of estimating the intensity field is generally referred to as a direction of arrival (DOA) estimation problem [1]. By thus specifying the data model, parametric methods can achieve much better recovery performance, both theoretically and in practice [1]. Stochastic Maximum Likelihood (SML) [1, 14, 15] is perhaps the most well-known parametric method. It uses explicit maximum likelihood expressions [16] to estimate the various parameters involved in the traditional point-source model [1, 14], namely the noise and sources power as well as the directions of arrival. Each parameter is estimated consistently and general theory of maximum likelihood guarantees efficient estimation as the number of samples grows to infinity. These strong theoretical guarantees and excellent empirical performance come however at the cost of very intense computation [1]. Moreover, the data model assumptions restrict its application to point sources, of which there must be fewer than the number of sensors, preventing its use in many applications. This is particularly true for radio-astronomy [6, 7], where the number of sources is typically far larger than the number of antennas forming the array. Moreover, the increased resolution of modern radio-interferometers [17, 18] permits celestial sources with spatial extent and complex shapes to be resolved, for which a point-source model is overly simplistic.
The present work, SiML, takes a maximum likelihood approach based on more general functional data model, which, in particular, allows potentially correlated sources with spatial extent and arbitrary shapes to be recovered. To this end, we leverage functional analysis tools [19, 20] and formulate the data in terms of an infinite-dimensional sampling operator [19] acting on the wavefield’s amplitude function, modelled as a complex Gaussian random function [21]. Based on this data model, we derive a joint maximum likelihood estimate for both the covariance kernel of this random function and the sensor noise power. As the optimisation problem elicits many solutions, we deploy the method of sieves [22, 23] as a means of restricting the optimisation problem to a lower dimensional subspace. For identifiability, we show that this subspace must have a smaller dimension than the total number of sensors in the array. A suitable subspace dimension is obtained by trading off between likelihood improvement and model complexity through the Bayesian Information Criterion (BIC) [24]. Simulations reveal that the subspace dimension acts as a regularisation parameter, increasing or decreasing with the SNR. For a known noise level, the resulting estimate of the covariance field is shown to be an unbiased, consistent and asymptotically efficient estimate of a particular oblique projection of the true covariance field. The method is computationally far more efficient than traditional SML and resilient to noise. Finally, we demonstrate by simulation that SiML obtains much better accuracy and contrast than spectral-based methods.

2 A Functional Data Model

To allow for the handling of very general sources, we introduce in this section a functional data model. Leveraging tools from functional analysis [19, 25], the sensing device can be modelled as an infinite dimensional sampling operator acting on an unknown random amplitude function (see [21] for an introduction to random functions). We first investigate the population version of the data model, where the covariance matrix of the instrument recordings is known, before presenting its empirical counterpart, where the covariance matrix is estimated from i.i.d. observations. More details on the modelling assumptions can be found in [1, 14, 26, 3].

2.1 Population Version

Consider an array of LL sensors with positions {𝒑1,,𝒑L}3\{\bm{p}_{1},\ldots,\bm{p}_{L}\}\subset\mathbb{R}^{3}. Assuming the emitting sources are in the far-field [3] of the sensor array, they can be thought of as lying on the unit sphere 𝕊2\mathbb{S}^{2}. To allow for arbitrary numbers of complex sources, we consider a notional continuous source field covering the entire sphere, with an associated amplitude function, describing for each direction 𝒓𝕊2\bm{r}\in\mathbb{S}^{2} the emission strength of the source field. In practice, source amplitudes fluctuate randomly [1, 14], and the amplitude function can be modelled as a complex random function 𝒮={S(𝒓):Ω,𝒓𝕊2}\mathcal{S}=\{S(\bm{r}):\Omega\rightarrow\mathbb{C},\bm{r}\in\mathbb{S}^{2}\}, where Ω\Omega is some probability space. More precisely, we assume 𝒮\mathcal{S} to be a Gaussian random function [21], i.e., that all its finite marginals have distribution:

[S(𝒓1),,S(𝒓n)]d𝒩n(0,𝒦n),n𝕊2,n,\left[S(\bm{r}_{1}),\cdots,S(\bm{r}_{n})\right]\stackrel{{\scriptstyle d}}{{\sim}}\mathbb{C}\mathcal{N}_{n}(0,\mathcal{K}_{\mathcal{I}_{n}}),\,\forall\,\mathcal{I}_{n}\subset\mathbb{S}^{2},\,\forall\,n\in\mathbb{N},

where 𝒩n\mathbb{C}\mathcal{N}_{n} denotes the nn-variate centrally symmetric, complex Gaussian distribution [27, 28], n:={𝒓1,,𝒓n}\mathcal{I}_{n}:=\{\bm{r}_{1},\ldots,\bm{r}_{n}\} and 𝒦nn×n\mathcal{K}_{\mathcal{I}_{n}}\in\mathbb{C}^{n\times n} is some valid covariance matrix depending on the set n{\mathcal{I}_{n}}.

From the Huygens-Fresnel principle [29], exciting the source field with a narrowband waveform of wavelength λ\lambda\in\mathbb{R} results in a diffracted wavefront, which, after travelling through an assumed homogeneous medium is recorded by the sensor array. In a far-field context, the Fraunhofer equation [29, 3] permits this wavefront at each sensor position 𝒑i3\bm{p}_{i}\in\mathbb{R}^{3} to be approximated by:

Y(𝒑i)\displaystyle Y(\bm{p}_{i}) =𝕊2S(𝒓)exp(j2πλ𝒓,𝒑i)𝑑𝒓+ni,\displaystyle=\int_{\mathbb{S}^{2}}\,S(\bm{r})\,\mbox{exp}\left(-j\frac{2\pi}{\lambda}\langle\bm{r},\bm{p}_{i}\rangle\right)\,d\bm{r}\;+\;n_{i}, (1)

where i=1,,L,i=1,\ldots,L, and 𝒏=[n1,,nL]\bm{n}=\left[n_{1},\ldots,n_{L}\right] is an additive white noise term capturing the inaccuracies in measurement of each sensor, distributed as [1]

𝒏d𝒩L(𝟎,σIL),σ>0.\bm{n}\stackrel{{\scriptstyle d}}{{\sim}}\mathbb{C}\mathcal{N}_{L}(\bm{0},\sigma I_{L}),\quad\sigma>0.

Noise across sensors is assumed to be identically and independently distributed and independent of the random amplitude function 𝒮\mathcal{S}.

We assume that every realisation, or sample function [21], sω:𝕊2s_{\omega}:~\mathbb{S}^{2}\rightarrow\mathbb{C} of 𝒮\mathcal{S} is an element of some Hilbert space =2(𝕊2,)\mathcal{H}=\mathcal{L}^{2}(\mathbb{S}^{2},\mathbb{C}) of finite-energy functions, and thus Eq. 1 can be written as:

Y(𝒑i)\displaystyle Y(\bm{p}_{i}) =S,ϕi+ni,i=1,,L,\displaystyle=\langle S,\phi_{i}\rangle\;+\;n_{i},\quad i=1,\ldots,L,

where ϕi(𝒓):=exp(j2π𝒓,𝒑i/λ),\phi_{i}(\bm{r}):=\mbox{exp}\left(j2\pi\langle\bm{r},\bm{p}_{i}\rangle/\lambda\right), which in turn can be re-written more compactly using an analysis operator [19] Φ:L\Phi^{\ast}:\mathcal{H}\rightarrow\mathbb{C}^{L}, mapping an element of \mathcal{H} to a finite number LL of measurements:

𝒀=[Y(𝒑1)Y(𝒑L)]=[S,ϕ1S,ϕL]+[n1nL]=ΦS+𝒏.\bm{Y}=\left[\begin{array}[]{c}Y(\bm{p}_{1})\\ \vdots\\ Y(\bm{p}_{L})\end{array}\right]=\left[\begin{array}[]{c}\langle S,\phi_{1}\rangle\\ \vdots\\ \langle S,\phi_{L}\rangle\end{array}\right]+\left[\begin{array}[]{c}n_{1}\\ \vdots\\ n_{L}\end{array}\right]=\Phi^{\ast}S+\bm{n}.

We call Φ\Phi^{\ast} the sampling operator [19] associated with the sensor array. As the sum of two independent centred complex Gaussian random vectors, the vector of measurements 𝒀\bm{Y} is also a centred complex Gaussian random vector with covariance matrix ΣL×L:\Sigma\in\mathbb{C}^{L\times L}:

(Σ)ij=𝕊2×𝕊2κ(𝒓,𝝆)ϕi(𝒓)ϕj(𝝆)𝑑𝒓𝑑𝝆+σδij,\displaystyle(\Sigma)_{ij}=\iint_{\mathbb{S}^{2}\times\mathbb{S}^{2}}\kappa(\bm{r},\bm{\rho})\,\phi^{\ast}_{i}(\bm{r})\phi_{j}(\bm{\rho})\,d\bm{r}d\bm{\rho}\;+\;\sigma\,\delta_{ij}, (2)

where i,j=1,,L,i,j=1,\ldots,L, δij\delta_{ij} denotes the Kronecker delta and κ:𝕊2×𝕊2\kappa:\mathbb{S}^{2}\times\mathbb{S}^{2}\rightarrow\mathbb{C} is the covariance kernel [21] of 𝒮\mathcal{S}:

κ(𝒓,𝝆):=𝔼[S(𝒓)S(𝝆)],𝒓,𝝆𝕊2.\kappa(\bm{r},\bm{\rho}):=\mathbb{E}\left[S(\bm{r})S^{\ast}(\bm{\rho})\right],\;\bm{r},\bm{\rho}\in\mathbb{S}^{2}.

Introducing the associated covariance operator 𝒯κ:\mathcal{T}_{\kappa}:\mathcal{H}\rightarrow\mathcal{H}:

(𝒯κf)(𝒓):=𝕊2κ(𝒓,𝝆)f(𝝆)𝑑𝝆,f,𝒓𝕊2,\displaystyle(\mathcal{T}_{\kappa}f)(\bm{r}):=\int_{\mathbb{S}^{2}}\kappa(\bm{r},\bm{\rho})f(\bm{\rho})d\bm{\rho},\quad f\in\mathcal{H},\,\bm{r}\in\mathbb{S}^{2},

we can again reformulate Eq. 2 in terms of the sampling operator Φ\Phi^{\ast} and its adjoint, called the synthesis operator [19], Φ:L\Phi:\mathbb{C}^{L}\rightarrow\mathcal{H}:

Σ=Φ𝒯κΦ+σIL.\Sigma=\Phi^{\ast}\mathcal{T}_{\kappa}\Phi\;+\;\sigma I_{L}. (3)

By analogy with the finite dimensional case [30], it is customary to write κ=vec(𝒯κ),\kappa=\mbox{vec}(\mathcal{T}_{\kappa}), where the vec()\mbox{vec}(\cdot) operator maps an infinite-dimensional linear operator onto its associated kernel representation. Because of the Gaussianity assumption, the covariance kernel κ\kappa (or equivalently the covariance operator 𝒯κ\mathcal{T}_{\kappa}) completely determines the distribution of 𝒮\mathcal{S}. Our goal is hence to leverage Eq. 3 in order to form an estimate of κ\kappa from the covariance matrix Σ\Sigma of the instrument recordings. Often the source field is assumed to be spatially uncorrelated, in which case the random function 𝒮\mathcal{S} is Gaussian white noise [21] and κ\kappa becomes diagonal. The diagonal part of κ\kappa

I(𝒓):=κ(𝒓,𝒓),𝒓𝕊2,I(\bm{r}):=\kappa(\bm{r},\bm{r}),\quad\bm{r}\in\mathbb{S}^{2},

is called the intensity function of the source field, of crucial interest in many array signal processing applications.

2.2 Empirical Version

In practice of course, the covariance matrix Σ\Sigma needs to be estimated from a finite number of i.i.d. observations of 𝒀\bm{Y}, say NN. Typically, the maximum likelihood estimate of Σ\Sigma is formed by Σ^=1Ni=1N𝒚i𝒚iH.\hat{\Sigma}=\frac{1}{N}\sum_{i=1}^{N}\bm{y}_{i}\bm{y}_{i}^{H}. It follows a LL-variate complex Wishart distribution [27, 31] with NN degrees of freedom and mean Σ\Sigma:

NΣ^d𝒲L(N,Σ).N\hat{\Sigma}\stackrel{{\scriptstyle d}}{{\sim}}\mathbb{C}\mathcal{W}_{L}(N,\Sigma). (4)

The density of a complex Wishart distribution can be found in [31]. In the next section, we use it to form the likelihood function of the data Σ^\hat{\Sigma} and derive maximum likelihood estimates of the covariance kernel κ\kappa and the noise level σ\sigma.

3 Sieved Maximum Likelihood

We now take the population and empirical data models Eqs. 3 and 4 and derive maximum likelihood estimates for κ\kappa and σ\sigma. The simpler case of known noise power, which allows for an insightful geometric interpretation of the maximum likelihood estimate in terms of projection operators, is presented first. That is then followed by the more general case given an unknown noise level.

3.1 A Constrained Log-Likelihood Maximisation Problem

The log-likelihood function for κ\kappa and σ\sigma given the sufficient statistic Σ^\hat{\Sigma} [32] can be written in terms of the density function of the complex Wishart distribution [31],

(κ,σ|Σ^)=Tr[(Φ𝒯κΦ+σIL)1Σ^]log|Φ𝒯κΦ+σIL|,\ell\left(\kappa,\sigma|\hat{\Sigma}\right)=-\mbox{Tr}\left[\left(\Phi^{\ast}\mathcal{T}_{\kappa}\Phi+\sigma I_{L}\right)^{-1}\hat{\Sigma}\right]-\log\left|\Phi^{\ast}\mathcal{T}_{\kappa}\Phi+\sigma I_{L}\right|, (5)

where the terms independent of κ\kappa and σ\sigma have been dropped. As σ>0\sigma>0, it is guaranteed that the matrix Φ𝒯κΦ+σIL\Phi^{\ast}\mathcal{T}_{\kappa}\Phi+\sigma I_{L} is invertible and that the log-likelihood function is hence well-defined. Maximum likelihood estimates for κ\kappa and σ\sigma are then obtained by maximising Eq. 5 with respect to κ2(𝕊2×𝕊2)\kappa\in\mathcal{L}^{2}(\mathbb{S}^{2}\times\mathbb{S}^{2}) and σ>0\sigma>0. Since the sampling operator Φ\Phi^{\ast} has finite rank and consequently a non-trivial kernel, the log-likelihood function admits infinitely many local maxima. Indeed, for f𝒩(Φ)f\in\mathcal{N}(\Phi^{\ast}), adding a kernel of the form111The tensor product \otimes is defined as (f¯f)g:=g,ff,f,g2(𝕊2,)\left(\bar{f}\otimes f\right)g:=\langle g,f\rangle f,\quad\forall f,g\in\mathcal{L}^{2}(\mathbb{S}^{2},\mathbb{C}). f¯f\bar{f}\otimes f to 𝒯κ\mathcal{T}_{\kappa} in (5) does not change the value of the log-likelihood function. We thus choose to impose a unique maximum by restricting the search space for κ\kappa to a lower dimensional subspace, and look for solutions in the range of some synthesis operator Ψ¯Ψ\bar{\Psi}\otimes\Psi, which will be specified in Section 3.4:

κ=(Ψ¯Ψ)vec(R)=i,j=1MRijψ¯jψi,𝒯κ=ΨRΨ,\kappa=\left(\bar{\Psi}\otimes\Psi\right)\mbox{vec}(R)=\sum_{i,j=1}^{M}R_{ij}\;\bar{\psi}_{j}\otimes\psi_{i},\;\Leftrightarrow\;\mathcal{T}_{\kappa}=\Psi R\Psi^{\ast},

where RM×MR\in\mathbb{C}^{M\times M} is a Hermitian symmetric matrix and Ψ:M\Psi^{\ast}:\mathcal{H}\rightarrow\mathbb{C}^{M}, Ψ:M\Psi:\mathbb{C}^{M}\rightarrow\mathcal{H} are the analysis and synthesis operators associated with the family of functions {ψ1,,ψM}\{\psi_{1},\ldots,\psi_{M}\}\subset\mathcal{H}. This regularisation of the likelihood problem by restricting the parameter space to a lower dimensional subspace is generally known as the method of sieves [22, 23]. The maximum likelihood estimates of RR and σ\sigma are then given by minimising the negative log-likelihood:

R^,σ^=argminRM2σ>0Tr[(GRGH+σIL)1Σ^]+log|GRGH+σIL|,\hat{R},\hat{\sigma}=\mbox{arg}\min_{\begin{subarray}{c}R\in\mathbb{C}^{M^{2}}\\ \sigma>0\end{subarray}}\mbox{Tr}\left[\left(GRG^{H}+\sigma I_{L}\right)^{-1}\hat{\Sigma}\right]+\log\left|GRG^{H}+\sigma I_{L}\right|, (6)

where G=ΦΨL×MG=\Phi^{\ast}\Psi\in\mathbb{C}^{L\times M} is the so-called Gram matrix [19], given by (G)ij=ψj,ϕi(G)_{ij}=\langle\psi_{j},\phi_{i}\rangle. For Eq. 6 to admit a unique solution, it is necessary to have at least as many measurements as unknowns. When the noise power is unknown a priori, this requires that M<LM<L. When the noise power is known, there is one less unknown, leading to MLM\leq L. This is however not a sufficient condition for identifiability, and we must further assume GG to be of full column-rank. If the latter condition is verified, we say that the two families of functions {ϕ1,,ϕL}\{\phi_{1},\ldots,\phi_{L}\} and {ψ1,,ψM}\{\psi_{1},\ldots,\psi_{M}\} are coherent with one another.

3.2 Estimation with Known Noise Power

Suppose the noise power σ\sigma is known. Then RR becomes the only variable in Eq. 6, and a solution can easily be obtained by cancelling the derivative. This yields

R^=GΣ~(G)H=G[Σ^σIL](G)H,\hat{R}=G^{\dagger}\tilde{\Sigma}\left(G^{\dagger}\right)^{H}=G^{\dagger}\left[\hat{\Sigma}-\sigma I_{L}\right]\left(G^{\dagger}\right)^{H},

where GL×MG^{\dagger}\in\mathbb{C}^{L\times M} is the left pseudo-inverse222The left pseudo-inverse exists since GG is assumed full-column rank. [33] of GG. Hence, when restricting the search space to (Ψ¯Ψ)\mathcal{R}(\bar{\Psi}\otimes\Psi), the maximum likelihood of κ\kappa is given by

κ^\displaystyle\hat{\kappa} =i,j=1MR^ijψ¯jψi\displaystyle=\sum_{i,j=1}^{M}\hat{R}_{ij}\;\bar{\psi}_{j}\otimes\psi_{i}
=(Ψ¯Ψ)vec(G[Σ^σIL](G)H)\displaystyle=\left(\bar{\Psi}\otimes\Psi\right)\mbox{vec}\left(G^{\dagger}\left[\hat{\Sigma}-\sigma I_{L}\right]\left(G^{\dagger}\right)^{H}\right)
=(Ψ¯Ψ)[G¯G](𝝇^σϵ),\displaystyle=\left(\bar{\Psi}\otimes\Psi\right)\left[\bar{G}^{\dagger}\otimes G^{\dagger}\right]\left(\hat{\bm{\varsigma}}-\sigma\bm{\epsilon}\right), (7)

with 𝝇^=vec(Σ^)L2\hat{\bm{\varsigma}}=\mbox{vec}(\hat{\Sigma})\in\mathbb{C}^{L^{2}} and ϵ=vec(IL)L2.\bm{\epsilon}=\mbox{vec}(I_{L})\in\mathbb{C}^{L^{2}}. The intensity function is then obtained by taking the diagonal part of κ^:\hat{\kappa}:

I^(𝒓)=i,j=1MR^ijψi(𝒓)ψ¯j(𝒓),𝒓𝕊2.\hat{I}(\bm{r})=\sum_{i,j=1}^{M}\hat{R}_{ij}\psi_{i}(\bm{r})\bar{\psi}_{j}(\bm{r}),\quad\forall\bm{r}\in\mathbb{S}^{2}.

Using properties of the tensor product and the vec operator, we can re-write Eq. 3 as 𝝇=(Φ¯Φ)κ+σϵ.\bm{\varsigma}=\left(\bar{\Phi}\otimes\Phi\right)^{\ast}\kappa\;+\;\sigma\bm{\epsilon}. Hence, since 𝔼[𝝇^]=𝝇\mathbb{E}[\hat{\bm{\varsigma}}]=~\bm{\varsigma}, Eq. 7 becomes on expectation

𝔼[κ^]=(Ψ¯Ψ)[G¯G](Φ¯Φ)κ.\mathbb{E}[\hat{\kappa}]=\left(\bar{\Psi}\otimes\Psi\right)\left[\bar{G}^{\dagger}\otimes G^{\dagger}\right]\left(\bar{\Phi}\otimes\Phi\right)^{\ast}\kappa.

For M=LM=L, GG is invertible and G=G1G^{\dagger}=G^{-1}, making (Ψ¯Ψ)[G¯1G1](Φ¯Φ)(\bar{\Psi}\otimes\Psi)[\bar{G}^{-1}\otimes G^{-1}](\bar{\Phi}\otimes\Phi)^{\ast} an oblique projection operator [19]. The operator (Ψ¯Ψ)[G¯1G1](\bar{\Psi}\otimes\Psi)[\bar{G}^{-1}\otimes G^{-1}] is indeed a right-inverse of (Φ¯Φ)(\bar{\Phi}\otimes\Phi)^{\ast}:

(Φ¯Φ)(Ψ¯Ψ)[G¯1G1]=ΦΨG1¯ΦΨG1=IL2.\left(\bar{\Phi}\otimes\Phi\right)^{\ast}\left(\bar{\Psi}\otimes\Psi\right)\left[\bar{G}^{-1}\otimes G^{-1}\right]=\overline{\Phi^{\ast}\Psi G^{-1}}\otimes\Phi^{\ast}\Psi G^{-1}=I_{L^{2}}. (8)

In the specific case where M=LM=L, the maximum likelihood estimate κ^\hat{\kappa} is hence an unbiased, consistent and asymptotically efficient estimator of the oblique projection of κ\kappa onto (Ψ¯Ψ)\mathcal{R}(\bar{\Psi}\otimes\Psi). When additionally setting Ψ=Φ\Psi=\Phi the projection becomes orthogonal.

3.3 Joint Estimation

Suppose now the noise power is unknown. We must hence minimise Eq. 6 with respect to both σ\sigma and RR. Using the result from theorem 1.1 of [16], we can write explicit solutions for the unique minimisers of Eq. 6:

σ^=Tr(Σ^GGΣ^)LM,R^=G[Σ^σ^I](G)H.\displaystyle\hat{\sigma}=\frac{\mbox{Tr}\left(\hat{\Sigma}-GG^{\dagger}\hat{\Sigma}\right)}{L-M},\qquad\hat{R}=G^{\dagger}\left[\hat{\Sigma}-\hat{\sigma}I\right]{\left(G^{\dagger}\right)}^{H}. (9)

Again, the constrained maximum likelihood estimate of κ\kappa is given by

κ^=(Ψ¯Ψ)[G¯G](𝝇^σ^ϵ),\hat{\kappa}=\left(\bar{\Psi}\otimes\Psi\right)\left[\bar{G}^{\dagger}\otimes G^{\dagger}\right]\left(\hat{\bm{\varsigma}}-\hat{\sigma}\bm{\epsilon}\right), (10)

with intensity function I^(𝒓)=i,j=1MR^ijψi(𝒓)ψ¯j(𝒓).\hat{I}(\bm{r})=\sum_{i,j=1}^{M}\hat{R}_{ij}\psi_{i}(\bm{r})\bar{\psi}_{j}(\bm{r}). This time, since M<LM<L the consistency condition Eq. 8 cannot be met, and 𝔼[κ^]\mathbb{E}[\hat{\kappa}] can no longer be interpreted as an oblique projection of κ\kappa. For values of MM comparable to LL though, the consistency condition should still hold approximately333More precisely, the consistency condition will hold on a subspace of L\mathbb{C}^{L} of dimension MM. , and this geometrical interpretation provides intuition.

3.4 On the choice of Ψ\Psi

We have thus far only required the synthesis operator Ψ\Psi to be identifiable, with the coherency condition requiring G=ΦΨG=\Phi^{\ast}\Psi to be full column-rank. This still leaves plenty of potential candidates. For practical purposes, we recommend taking Ψ=ΦW\Psi=\Phi W where WL×MW\in\mathbb{C}^{L\times M} is a tall matrix, with columns containing the first MM eigenvectors of Σ^\hat{\Sigma} (assuming eigenvalues sorted in descending order). Such a choice presents numerous advantages. First, since (Φ)=𝒩(Φ)\mathcal{R}(\Phi)^{\perp}=\mathcal{N}(\Phi^{\ast}), the instrument can only sense functions within the range of Φ\Phi, and it is hence natural to choose (Ψ)=(Φ)\mathcal{R}(\Psi)=\mathcal{R}(\Phi). This canonical choice moreover yields an analytically computable Gram matrix GG. Indeed, we have G=ΦΦW=HWG=\Phi^{\ast}\Phi W=HW, where HL×LH\in\mathbb{C}^{L\times L} is given by (see of [7, Chapter 4 section 1.1]):

(H)ij=4πsinc(2π𝒑i𝒑j2/λ),i,j=1,,L.(H)_{ij}=~4\pi\;\mbox{sinc}(2\pi\|\bm{p}_{i}-\bm{p}_{j}\|_{2}/\lambda),\quad i,j=1,\ldots,L.

Finally, by choosing the columns of WW as the first MM eigenvectors of Σ^\hat{\Sigma}, MM acts as a regularisation parameter. Indeed, the eigenvectors associated to the smallest eigenvalues of Σ^\hat{\Sigma} are usually the most polluted by noise. Hence, truncating to the MM largest eigenvalues reduces the contribution of the noise in the final estimate (see Figs. 2f, 2g and 2h). Moreover, small values of MM will increase the chances of (GHG)M×M(G^{H}G)\in\mathbb{C}^{M\times M} in the left pseudo-inverse G=(GHG)1GHG^{\dagger}=(G^{H}G)^{-1}G^{H} being well-conditioned, thus improving the overall numerical stability of the algorithm. Suitable values of MM can be obtained by minimising the Bayesian Information Criterion (BIC) [24], often used in model selection: BIC(M)=2^M+2M2log(L),BIC(M)=-2\hat{\ell}_{M}+2M^{2}\log(L), where ^M\hat{\ell}_{M} is the maximised log-likelihood function for a specific choice of MM. Example of a BIC profile and evolution of the BIC-selected MM with the signal-to-noise ratio are depicted in Fig. 1.

3.5 Simulation Results

Fig. 2 compares the performance of the proposed Sieved Maximum Likelihood (SiML) method in a radio astronomy setup to three popular spectral-based methods, namely Matched Beamforming (MB), Maximum Variance Distortionless Response (MVDR) and the Adapted Angular Response (AAR) [34]. For this experiment, we generated randomly a layout L=300L=300 antennas and simulated N=2000N=2000 random measurements from the ground truth intensity field Fig. 2a. Furthermore, we considered two metrics to assess the quality of the recovered images: the traditional relative Mean Squared Error (MSE) and the Root Mean Squared (RMS) metric, which measures the contrast of an image by computing its standard deviation over all pixels. The simulations reveal that the SiML outperforms all the traditional algorithms for the considered SNR range in both metrics, except for large SNRs where MVDR exhibits a slightly better contrast. As for the traditional SML method, SiML performs particularly well for challenging scenarios with very low SNR.

4 Conclusion

SiML generalises the traditional SML method to a wider class of signals, encompassing arbitrarily shaped, possibly correlated, sources of which there may be more than the number of sensors. The method is numerically stable and admits a nice geometrical interpretation in the case of known noise power. Simulations revealed its superiority with respect to state-of-the-art subspace-based methods, both in terms of accuracy and contrast. Finally, the tensor product structure in Eq. 10 makes the estimate κ^\hat{\kappa} very efficient to compute. This is in contrast to traditional SML, which requires minimising a highly non-linear multi-dimensional function [1].

Refer to caption
(a)
Refer to caption
(b)

ß

Fig. 1: Optimal choice of the dimension MM based on the Bayesian Information Criterion (BIC).
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Refer to caption
(i)
Fig. 2: Comparison between sieved maximum likelihood method and various steered-response power methods (MB, MVDR, AAR). The parameters of the experiment were set to: L=300L=300, N=2000N=2000, SNR=5dBSNR=5\,\mbox{dB}.

References

  • [1] Hamid Krim and Mats Viberg, “Two decades of array signal processing research: the parametric approach,” IEEE Signal processing magazine, vol. 13, no. 4, pp. 67–94, 1996.
  • [2] Robert J Mailloux, Phased array antenna handbook, vol. 2, Artech House Boston, 2005.
  • [3] Don H Johnson and Dan E Dudgeon, Array signal processing: concepts and techniques, Simon & Schuster, 1992.
  • [4] Michael Brandstein and Darren Ward, Microphone arrays: signal processing techniques and applications, Springer Science & Business Media, 2013.
  • [5] Jacob Benesty, Jingdong Chen, and Yiteng Huang, Microphone array signal processing, vol. 1, Springer Science & Business Media, 2008.
  • [6] A Richard Thompson, James M Moran, and George W Swenson Jr, Interferometry and synthesis in radio astronomy, John Wiley & Sons, 2008.
  • [7] Matthieu Simeoni, “Towards more accurate and efficient beamformed radio interferometry imaging,” M.S. thesis, EPFL, Spring 2015.
  • [8] Simon Haykin, “Array signal processing,” Englewood Cliffs, NJ, Prentice-Hall, Inc., 1985, 493 p. For individual items see A85-43961 to A85-43963., vol. 1, 1985.
  • [9] Lal Chand Godara, “Application of antenna arrays to mobile communications. ii. beam-forming and direction-of-arrival considerations,” Proceedings of the IEEE, vol. 85, no. 8, pp. 1195–1245, 1997.
  • [10] Arogyaswami J Paulraj and Constantinos B Papadias, “Space-time processing for wireless communications,” IEEE Signal Processing Magazine, vol. 14, no. 6, pp. 49–83, 1997.
  • [11] P. Hurley and M. Simeoni, “Flexibeam: analytic spatial filtering by beamforming,” in International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, March 2016.
  • [12] Zhi-Pei Liang and Paul C Lauterbur, Principles of magnetic resonance imaging: a signal processing perspective, The Institute of Electrical and Electronics Engineers Press, 2000.
  • [13] Boaz Rafaely, Fundamentals of spherical array processing, vol. 8, Springer, 2015.
  • [14] Petre Stoica, Björn Ottersten, Mats Viberg, and Randolph L Moses, “Maximum likelihood array processing for stochastic coherent sources,” IEEE Transactions on Signal Processing, vol. 44, no. 1, pp. 96–105, 1996.
  • [15] Petre Stoica, Erik G Larsson, and Alex B Gershman, “The stochastic crb for array processing: A textbook derivation,” IEEE Signal Processing Letters, vol. 8, no. 5, pp. 148–150, 2001.
  • [16] Petre Stoica and Arye Nehorai, “On the concentrated stochastic likelihood function in array signal processing,” Circuits, Systems and Signal Processing, vol. 14, no. 5, pp. 669–674, 1995.
  • [17] MP Van Haarlem, MW Wise, AW Gunst, George Heald, JP McKean, JWT Hessels, AG De Bruyn, Ronald Nijboer, John Swinbank, Richard Fallows, et al., “Lofar: The low-frequency array,” Astronomy & Astrophysics, vol. 556, pp. A2, 2013.
  • [18] Peter E Dewdney, Peter J Hall, Richard T Schilizzi, and T Joseph LW Lazio, “The square kilometre array,” Proceedings of the IEEE, vol. 97, no. 8, pp. 1482–1496, 2009.
  • [19] Martin Vetterli, Jelena Kovačević, and Vivek K Goyal, Foundations of signal processing, Cambridge University Press, 2014.
  • [20] James O Ramsay, Functional data analysis, Wiley Online Library, 2006.
  • [21] Mikhail Lifshits, “Lectures on gaussian processes,” in Lectures on Gaussian Processes, pp. 1–117. Springer, 2012.
  • [22] Ulf Grenander and Grenander Ulf, “Abstract inference,” Tech. Rep., 1981.
  • [23] Stuart Geman and Chii-Ruey Hwang, “Nonparametric maximum likelihood estimation by the method of sieves,” The Annals of Statistics, pp. 401–414, 1982.
  • [24] Harish S Bhat and Nitesh Kumar, “On the derivation of the bayesian information criterion,” School of Natural Sciences, University of California, 2010.
  • [25] James O Ramsay and Bernard W Silverman, Applied functional data analysis: methods and case studies, vol. 77, Citeseer, 2002.
  • [26] Björn Ottersten, Peter Stoica, and Richard Roy, “Covariance matching estimation techniques for array signal processing applications,” Digital Signal Processing, vol. 8, no. 3, pp. 185–210, 1998.
  • [27] Nathaniel R Goodman, “Statistical analysis based on a certain multivariate complex gaussian distribution (an introduction),” The Annals of mathematical statistics, vol. 34, no. 1, pp. 152–177, 1963.
  • [28] Robert G Gallager, Principles of digital communication, vol. 1, Cambridge University Press Cambridge, UK:, 2008.
  • [29] T Douglas Mast, “Fresnel approximations for acoustic fields of rectangularly symmetric sources,” The Journal of the Acoustical Society of America, vol. 121, no. 6, pp. 3311–3322, 2007.
  • [30] KG Jinadasa, “Applications of the matrix operators vech and vec,” Linear Algebra and its Applications, vol. 101, pp. 73–79, 1988.
  • [31] D Maiwald and D Kraus, “Calculation of moments of complex wishart and complex inverse wishart distributed matrices,” IEE Proceedings-Radar, Sonar and Navigation, vol. 147, no. 4, pp. 162–168, 2000.
  • [32] Victor M Panaretos, “Statistics for mathematicians,” .
  • [33] Heinz Werner Engl, Martin Hanke, and Andreas Neubauer, Regularization of inverse problems, vol. 375, Springer Science & Business Media, 1996.
  • [34] Alle-Jan van der Veen and Stefan J Wijnholds, “Signal processing tools for radio astronomy,” in Handbook of Signal Processing Systems, pp. 421–463. Springer, 2013.