This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

An Efficient High-Dimensional Sparse Fourier Transform111This manuscript has been submitted to IEEE Transactions on Aerospace and Electronic Systems for reviewing on Sept. 1st, 2016.

Shaogang Wang, , Vishal M. Patel, , and Athina Petropulu Shaogang Wang, Vishal M. Patel and Athina Petropulu are with the department of Electrical and Computer Engineering at Rutgers University, Piscataway, NJ USA. Email: {shaogang.wang, vishal.m.patel, athinap}@rutgers.edu.
Abstract

We propose RSFT, which is an extension of the one dimensional Sparse Fourier Transform algorithm to higher dimensions in a way that it can be applied to real, noisy data. The RSFT allows for off-grid frequencies. Furthermore, by incorporating Neyman-Pearson detection, the frequency detection stages in RSFT do not require knowledge of the exact sparsity of the signal and are more robust to noise. We analyze the asymptotic performance of RSFT, and study the computational complexity versus the worst case signal SNR tradeoff. We show that by choosing the proper parameters, the optimal tradeoff can be achieved. We discuss the application of RSFT on short range ubiquitous radar signal processing, and demonstrate its feasibility via simulations.

Index Terms:
Array signal processing, sparse Fourier transform, detection and estimation, radar signal processing.

I Introduction

High dimensional FFT (N-D FFT) is used in many applications in which a multidimensional Discrete Fourier Transform (DFT) is needed, such as radar and imaging. However, its complexity increases with an increase in dimension, leading to costly hardware and low speed of system reaction. The recently developed Sparse Fourier Transform (SFT) [1, 2, 3, 4, 5, 6], designed for signals that have a small number of frequencies enjoys low complexity, and thus is ideally suited in big data scenarios [7, 8, 9]. One such application is radar target detection. The radar returns are typically sparse in the target space, i.e., the number of targets is far less than the number of resolutions cells. This observation motivates the application of SFT in radar signal processing.

The current literature on multi-dimensional extensions of the SFT include [10, 5, 3, 6]. Those works mainly address sample complexity, i.e., using the least number of time domain samples to reconstruct the signal frequencies. In order to detect the significant frequencies in an approximately sparse settings, i.e., signal corrupted by additive noise, the aforementioned methods assume knowing the exact sparsity, and compare the frequency amplitude with a predefined threshold. However, in many real applications, the exact signal sparsity may be either unknown or subject to change. For example, in the radar case, the number of targets to be detected is typically unknown and usually varies from time to time. Also, setting up an ideal threshold for detection in noisy cases is not trivial, since it relates to the tradeoff between probability of detection and false alarm rate. However, those issues have not been studied in the SFT literature.

One of the constrains in the aforementioned SFT algorithms is the assumption that the signal discrete frequencies are all on-grid. In reality, however, the signal frequencies, when discretized and depending on the grid size can fall between grid points. The consequence of off-grid frequencies is leakage to other frequency bins, which essentially reduces the sparsity of the signal. To refine the estimation of off-grid frequencies starting from the initial SFT-based estimates,[9] proposed a gradient descent method. However, the method of [9] has to enumerate all possible directions of the gradient for each frequency and compute the approximation error for each guessed direction, which increases the computational complexity. Moreover, like the other aforementioned SFT methods, the thresholding for frequency detection is not clear in [9]. The off-grid frequency problem in the context of SFT was also studied in [11], where it was assumed that in the frequency domain the signal and the noise are well separated by predefined gaps. However, this is a restrictive assumption that limits the applicability of the work.

In this paper we setup a SFT-based framework for sparse signal detection in a high dimensional frequency domain and propose a new algorithm, namely, the Realistic Sparse Fourier Transform (RSFT) which addresses the shortcomings discussed above. This paper makes the following contributions.

  1. 1.

    The RSFT algorithm does not require knowledge of the number of frequencies to be estimated. Also, it does not need the frequencies to be on-grid and does not require signal and noise to be separated in the frequency domain. Our method reduces the leakage effect from off-grid frequencies by applying a window on the input time domain data (see Section III-A for details). The design of this window trades off frequency resolution, leakage suppression and computational efficiency. We shall point out that, unlike the work of [9] that recovers the exact off-grid frequency locations, our work aims to recover grided locations of off-grid frequencies with less amount of computation.

  2. 2.

    We extend the RSFT into an arbitrary fixed high dimension, so that it can replace the N-D FFT in sparse settings and thus enable computational savings.

  3. 3.

    We put RSFT into a Neyman-Pearson (NP) detection framework. Based on the signal model and other design specifications, we give the (asymptoticly) optimal thresholds for the two detection stages of the RSFT (see Section IV for details). Since the output of the first stage of detection serves as the input of the second stage, the two detection stages are interconnected. The detection thresholds are jointly found by formulating and solving an optimization problem, with its objective function minimizing the worst case Signal to Noise Ratio (SNR) (hence the system is more sensitive to weak signal), and its constrains connecting probability of detection and false alarm rate for both two stages.

  4. 4.

    We provide a quantitive measure of tradeoff between computational complexity and worst case signal SNR for systems that use RSFT, which serves as a concrete design reference for system engineers.

  5. 5.

    We investigate the use of RSFT in multi-dimensional radar signal processing.

A closely related technique to SFT is compressed sensing (CS). Compressed sensing-based methods recover signals by exploiting their sparse features [12]. The application of CS methods in MIMO radar is discussed intensively in [13, 14]. At a high level, CS differs from SFT in that it assumes the signal can be sparsely represented by an overdetermined dictionary, and formulates the problem as an optimization problem with sparsity constraints (l0l_{0} or l1l_{1} norm). This problem is usually solved by convex optimization, which runs in polynomial time in NN for an NN-dimensional signal. On the other hand, the SFT method finds an (approximately) sparse Fourier representation of a signal in a mean square error sense (l2l_{2} norm). The sample and the computational complexity of SFT is sub-linear to NN, for a wide range of NN-dimensional signals [6].

A preliminary version of this work appeared in [15]. A detailed analysis of the RSFT algorithm as well as extensive experimental results are extensions to [15].

I-A Notation

We use lower-case (upper-case) bold letters to denote vectors (matrices). ()T(\cdot)^{T} and ()H(\cdot)^{H} respectively denote the transpose and conjugate transpose of a matrix or a vector. \|\cdot\| is Euclidean norm for a vector. [𝐚]i[\mathbf{a}]_{i} is the ithi_{th} element of vector 𝐚\mathbf{a}. All operations on indices in this paper are taken modulo NN denoted by []N[\cdot]_{N}. We use \lfloor\cdot\rfloor to denote rounding to floor. [S][S] refers to the set of indices {0,,S1}\{0,...,S-1\}, and [S]a[S]\setminus a is for eliminating element aa from set [S][S]. We use {0,1}B\{0,1\}^{B} to denote the set of BB-dimensional binary vectors. We use diag()\mathop{\mathrm{diag}}\nolimits(\cdot) to denote forming a diagonal matrix from a vector and use 𝔼{}\mathbb{E}\{\cdot\} to denote expectation. The DFT of signal 𝐬{\mathbf{s}} is denoted as 𝐬^\hat{{\mathbf{s}}}. We also assume that the signal length in each dimension is an integer power of 2.

This paper is organized as follows. A brief background on the SFT algorithm is given in Section II. Details of the proposed RSFT algorithm are given in Section III. Section IV presents the derivation of the optimal threshold design for the RSFT algorithm. Then in Section V we provide some numerical results to verify the theoretical findings. An application of the RSFT algorithm in radar signal processing is presented in Section VI and finally, concluding remarks are made in Section VII.

II Preliminaries

II-A Basic Techniques

As opposed to the FFT which computes the coefficients of all NN discrete frequency components of an NN-sample long signal, the SFT[2] computes only the KK frequency components of a KK-sparse signal. Before outlining the SFT algorithm we provide some key definitions and properties, which is extracted and reformulated based on [2] .

Definition 1.

(Permutation): Define the transform Pσ,τP_{\sigma,\tau} such that, given 𝐱N,σ,τ[N]\mathbf{x}\in\mathbb{C}^{N},\sigma,\tau\in[N], and σ\sigma invertible

σσ1=[1]N.\sigma\sigma^{-1}=[1]_{N}. (1)

Then, the following transformation is called permutation

(Pσ,τx)i=xσi+τ.(P_{\sigma,\tau}x)_{i}=x_{\sigma i+\tau}. (2)

The permutation has the following property.

Property 1.

A modular reordering of the data in time domain results in a modular dilation and phase rotation in the frequency domain, i.e.,

(Pσ,τx)^σi=x^iejτ2πN.\widehat{(P_{\sigma,\tau}x)}_{\sigma i}=\hat{x}_{i}e^{-j\tau\frac{2\pi}{N}}. (3)
Definition 2.

(Aliasing): Let 𝐱N,𝐲B\mathbf{x}\in\mathbb{C}^{N},\mathbf{y}\in\mathbb{C}^{B}, with N,BN,B powers of 2, and B<NB<N. For L=N/BL={N}/{B}, a time-domain aliased version of x is defined as

yi=j=0L1xi+Bj,i[B].y_{i}=\sum_{j=0}^{L-1}x_{i+Bj},\quad i\in[B]. (4)
Property 2.

Aliasing in time domain results in downsampling in frequency domain, i.e.,

y^i=x^iL.\hat{y}_{i}=\hat{x}_{iL}. (5)
Definition 3.

(Mapping): Let i,σ[N]i,\sigma\in[N], where σ\sigma satisfies (1). We define the mapping (i,σ)\mathcal{M}(i,\sigma) such that

[B](i,σ)BN[iσ]N.[B]\ni\mathcal{M}(i,\sigma)\equiv\lfloor\frac{B}{N}[i\sigma]_{N}\rfloor. (6)
Definition 4.

(Reverse-mapping): Let σ1[N]\sigma^{-1}\in[N], σ1\sigma^{-1} satisfies (1), and j[B]j\in[B]. Define (j,σ1)\mathcal{R}(j,\sigma^{-1}) a reverse-mapping such that

(j,σ1){[σ1u]Nu𝕊},\begin{split}&\mathcal{R}(j,\sigma^{-1})\equiv\{[\sigma^{-1}u]_{N}\mid u\in\mathbb{S}\},\end{split} (7)

where

𝕊{v[N]jNBv<(j+1)NB]}.\begin{split}\mathbb{S}\equiv\{v\in[N]\mid j\frac{N}{B}\leq v<(j+1)\frac{N}{B}]\}.\end{split} (8)

II-B SFT Algorithm

At a high level, the SFT algorithm in [2] runs two loops, namely the Location loop and the Estimation loop. The former finds the indices of the KK most significant frequencies from the input signal, while the latter estimates the corresponding Fourier coefficients. Here, we emphasize on Location more than Estimation, since the former is more relevant to the radar application that we consider. The Location step provides frequency locations, which in the radar case is directly related to target parameters.

In the Location loop, a permutation procedure reorders the input data in the time domain, causing the frequencies to also reorder. The permutation causes closely spaced frequencies to appear in well separated locations with high probability. Then, a flat-window[2] is applied on the permuted signal for the purpose of extending a single frequency into a (nearly) boxcar, for a reason that will become apparent in the following. The windowed data are aliased, as in Definition 2. The frequency domain equivalent of this aliasing is undersampling by N/BN/B (see Property 2). The flat-window used at the previous step ensures that no peaks are lost due to the effective undersampling in the frequency domain. After this stage, a FFT of length BB is employed.

The permutation and the aliasing procedure effectively map the signal frequencies from NN-dimensional space into a reduced BB-dimensional space, where the first-stage-detection procedure finds the significant frequencies’ peaks, and then their indices are reverse mapped into the original NN-dimensional frequency space. However, the reverse mapping yields not only the true location of the significant frequencies, but also N/B{N}/{B} ambiguous locations for each frequency. To remove the ambiguity, multiple iterations of Location with randomized permutation are performed. Finally, the second-stage-detection procedure locates the KK most significant frequencies from the accumulated data for each iteration. More details about the SFT algorithm can be found in [2].

III The RSFT Algorithm

In this section we address some problems that have not been considered in the original SFT algorithm, namely the leakage from off-grid frequencies and optimal detection threshold design for both detection stages. Also, we extend the RSFT into arbitrary fixed high dimension.

III-A Leakage Suppression for Off-grid Frequencies

In real world applications, the frequencies are continuous and can take any value in [0,2π)[0,2\pi). When fitting a grid on these frequencies, leakage occurs from off-grid frequencies, which can diminish the sparsity of the signal. As the leakage due to strong frequency components can mask the contributions of weak frequency components, it becomes difficult to determine the frequency domain peaks after permutation. (see Fig. 1 (c)). To address this problem, we propose to multiply the received time domain signal with a window before permutation. We call this procedure pre-permutation windowing. The idea is to confine the leakage within a finite number of frequency bins, as illustrated in Fig. 1.

The choice of the pre-permutation window is determined by the required resolution, computational complexity, and degree of leakage suppression. Specifically, the side-lobe level should be lower than the noise level after windowing (see Fig. 4). However, the larger the attenuation of the side-lobes, the wider the main lobe would be, thus lowering the frequency resolution. Meanwhile, a broader main lobe results in increased computational load, which will be discussed in Section IV-D.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 1: Effect of pre-permutation windowing. The signal contains two significant frequency components, one of which is 15dB15dB stronger than the other. A Dolph-Chebyshev window is applied to the time-domain signal. Windowed signal after permutation appears sparser in the frequency domain as compared to the permuted signal without windowing. (a) Spectrum of signal without windowing. (b) Spectrum of windowed signal. (c) Spectrum of signal without windowing after permutation. (d) Spectrum of windowed signal after permutation.

III-B Signal Detection Without Knowing the Exact Sparsity and Optimal Threshold Design

In the SFT, detection of the significant frequencies is needed in two stages. With knowing the number of the significant frequencies and assuming they are all on-grid, the detection of the signal can be accomplished by finding the KK highest spectral amplitude values.

In reality however, we usually do not have the knowledge of the exact sparsity, i.e., KK. Moreover, even if we knew KK, due to the leakage caused by the off-grid signals, the KK highest spectral peaks might not be the correct representation of the signal frequencies. Finally, the additive noise could generate false alarms, which would add more difficulties in signal detection.

In order to solve the detection problem, we propose to use NP detection in the two stages of detection, which does not require knowing KK. However, the optimal thresholds still rely on the number of significant frequencies as well as their SNR. In light of this, we provide an optimal design based on a bound for the sparsity and signal SNR level. The optimal threshold design is presented in Section IV.

III-C High Dimensional Extensions

In the following, we elaborate on the high dimensional extension of the RSFT for its main stages.

III-C1 Windowing

In the pre-permutation windowing and the flat-windowing stages, the window for each dimension is designed separately. After that, the high dimension widow is generated by combining each 1-D window. For instance, in the 2-D case, assuming that 𝐰x\mathbf{w}_{x} and 𝐰y\mathbf{w}_{y} are the two windows in the xx and yy dimension, respectively, a 2D window can be computed as

𝐖xy=𝐰x𝐰yH.\mathbf{W}_{xy}=\mathbf{w}_{x}\mathbf{w}_{y}^{H}. (9)

Fig. 2 shows a compound 2-D window which is a combination of a 64-point and a 1024-point Dolph-Chebyshev window, both of which has 60dB60dB attenuation for the side-lobes in frequency domain. A Dolph-Chebyshev window allows us to trade off frequency resolution and side-lobe attenuation easly. Other windows sharing the same flexibility includes Gaussian window, Kaiser-Bessel window, Blackman-Harris window, etc., see [16] for detail. We apply those windows on the data by point-wise multiplications.

Refer to caption
(a)
Refer to caption
(b)
Figure 2: Compound Window in 2-D. (a) Top: a 6464-points Dolph-Chebyshev window; bottom: a 10241024-points Dolph-Chebyshev window. (b) The 2D window.

III-C2 Permutation

The permutation parameters are generated for each dimension in a random way according to (2). Then, we carry the permutation on each dimension sequentially. An example for the 2-D case is illustrated in Fig. 3.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 3: Permutation and Aliasing in 2D. (a) Original 2D data forms a 4×84\times 8 matrix. (b) Permutation in xx-dimension, σx=3,τx=0\sigma_{x}=3,\tau_{x}=0. (c) Permutation in yy-dimension, σy=3,τy=0\sigma_{y}=3,\tau_{y}=0. After permutation, data is divided into four 2×42\times 4 sub-matrices. (d) Aliasing by adding sub-matrices from (c).

III-C3 Aliasing

The aliasing stage compresses the high dimensional data into much smaller size. In 2-D, as shown in Fig. 3, a periodic extension of the Nx×NyN_{x}\times N_{y} data matrix is created with period BxB_{x} in the xx dimension and ByB_{y} in the yy dimension, with Bx<<NxB_{x}<<N_{x} and By<<NyB_{y}<<N_{y}, and the basic period, i.e., Bx×ByB_{x}\times B_{y} is extracted.

III-C4 First-stage-detection and reverse-mapping

We carry first stage detection after taking the square of magnitude of N-D FFT on the aliased data. Each data point is then compared with a pre-determined threshold, and for those passing the thresholds, their indices are reversed map to the original space. The combination of the reverse mapped indices from each dimension provides the tentative locations of the original frequency components.

III-C5 Accumulation and second-stage-detection

The accumulation stage collects the tentative frequency locations found in the reverse mapping for each iteration, and the number of occurrences for each location is calculated after running over TT iterations. The second stage detection finds indices in each dimension for those whose number of occurrence pass the threshold of this stage.

Based on the above discussion, we summarize the RSFT in Algorithm 1. Comparing to the original SFT, we add pre-permutation windowing process on the input data and incorporate NP detection in both stages of detection.

Depending on the application, the iterations can be applied on the same input data with different permutations. Alternatively, each iteration could process different segments (see the signal model in (10)) of input data as indicated in Algorithm 1, and it can effectively reduce the variance of the estimation via RSFT, provided that the signal is stationary.

The Estimation procedure in the RSFT is similar to the original Estimation procedure, except that the recovered Fourier coefficients should be divided by the spectrum of the pre-permutation window, so that the distortion due to pre-permutation windowing is compensated. Moreover, using Location is sufficient for our radar application, for it can provide the location of significant frequencies, which is directly related to the parameters to be estimated. We also set τ=0\tau=0 in each permutation, since the random phase rotation does not affect the performance of a detector after taking magnitude of the signal in the intermediate stage.

Algorithm 1 RSFT algorithm

Input: complex signal 𝐫\mathbf{r} in any fixed dimension
Output: 𝐨\mathbf{o}, sparse frequency locations of input signal

1:procedure RSFT(𝐫\mathbf{r})
2:  Generate a set of σ\sigma randomly for each dimension
3:  𝐚¯0\bar{\mathbf{a}}\leftarrow 0
4:  for i0i\leftarrow 0 to TT do
5:   Pre-Permutation Windowing: 𝐲𝐖𝐫\mathbf{y}\leftarrow\mathbf{W}\mathbf{r}
6:   Permutation: 𝐩\mathbf{p} \leftarrow Pσ,0𝐲P_{\sigma,0}\mathbf{y}
7:   Flat-windowing: 𝐳\mathbf{z} \leftarrow 𝐖¯𝐩\overline{\mathbf{W}}\mathbf{p}
8:   Aliasing: 𝐟Aliasing(𝐳\mathbf{f}\leftarrow\mathop{\mathrm{Aliasing}}(\mathbf{z})
9:   N-D FFT: 𝐟^\hat{\mathbf{f}} \leftarrow NDFFT(𝐟)\mathop{\mathrm{NDFFT}}(\mathbf{f})
10:   First-stage-detection: 𝐜NPdet1(|𝐟^|2\mathbf{c}\leftarrow\mathop{\mathrm{NPdet1}}(|\hat{\mathbf{f}}|^{2})
11:   Reverse-mapping: 𝐚i\mathbf{a}_{i} \leftarrow Reverse(𝐜)\mathop{\mathrm{Reverse}}(\mathbf{c})
12:   Accumulation: 𝐚¯𝐚¯+𝐚i\bar{\mathbf{a}}\leftarrow\bar{\mathbf{a}}+\mathbf{a}_{i}
13:  end for
14:  Second-stage-detection: 𝐨NPdet2(𝐚¯)\mathbf{o}\leftarrow\mathop{\mathrm{NPdet2}}(\bar{\mathbf{a}})
15:  return 𝐨\mathbf{o}
16:end procedure

IV Optimal Thresholds Design in the RSFT

The main challenge in implementing the RSFT is to decide the thresholds in the two stages of NP detection. The applying of NP detection in the RSFT is not a straitforward extension on the SFT, in that the two stages are inter-connected, thus need to be jointly studied. In the following, we will show that the two asymptotically optimal thresholds are jointly founded with an optimization process. The analysis is carried in 1-D, while the generalization to high dimension is straitforward.

IV-A Signal Model and Problem Formulation

We model the signal in continuous time domain as a superposition of KK sinusoids as well as additive white noise. We then sample the signal uniformly both in I and Q channels with a sampling frequency above the Nyqvist rate. Assume the total sampling time is divided into TT consecutive equal length segments, each of which contains NN samples, and K<<NK<<N (i.e., the signal is sparse in frequency domain). Then for each time segment, i.e., s[T]s\in[T], we have

𝐫s=i[K]bi,s𝐯(ωi)+𝐧s,\mathbf{r}_{s}=\sum_{i\in[K]}b_{i,s}\mathbf{v}(\omega_{i})+\mathbf{n}_{s}, (10)

where 𝐯(ωi)\mathbf{v}(\omega_{i}) denotes for the ithi_{th} complex sinusoid, with ωi[0,2π)\omega_{i}\in[0,2\pi) as its frequency, i.e.,

𝐯(ωi)=[1ejωiej(N1)ωi]T.\mathbf{v}(\omega_{i})=[1\quad e^{j\omega_{i}}\;\cdots\;e^{j(N-1)\omega_{i}}]^{T}. (11)

We further assume that ωi\omega_{i} is unknown deterministic quantity and is constant during the whole process, while the complex amplitude of the sinusoid, i.e., bi,sb_{i,s} takes random value for each segment. More specifically, we model bi,sb_{i,s} as a circularly symmetric Gaussian process with the distribution bi,s𝒞𝒩(0,σbi2)b_{i,s}\sim\mathcal{CN}(0,\sigma_{bi}^{2}). Likewise, the noise 𝐧s\mathbf{n}_{s} is distributed as 𝐧s𝒞𝒩(𝟎,σn2𝐈)\mathbf{n}_{s}\sim\mathcal{CN}(\mathbf{0},\sigma_{n}^{2}\mathbf{I}), where 𝟎\mathbf{0} is NN-dimensional zero vector, and 𝐈N×N\mathbf{I}\in\mathbb{R}^{N\times N} is unit matrix. We also assume each sinusoid and the noise are uncorrelated. In addition, the neighboring sinusoids are resolvable in the frequency domain, i.e., the frequency spacing of neighboring sinusoids is greater than ηm2πN\eta_{m}\frac{2\pi}{N}, where ηm\eta_{m} is the 6.0-dB bandwidth[16] of a window that applies on 𝐫s\mathbf{r}_{s}. Note the signal model in (10) is also commonly used in array signal processing literature for Uniform Linear Array (ULA) settings (See e.g., [17]), in which the time samples in each segment is replaced by spatial samples from array elements, and a sample segment is referred as a time snapshot.

We want to detect and estimate each ωi\omega_{i} from the input signal. From a non-parametric and data-independent perspective, this is a classic spectral analysis and detection problem that can be solved by any FFT-based spectrum estimation method, for example, the Bartlett method followed by a NP detection procedure (see Appendix A). In that case, the FFT computes the signal spectrum on NN frequency bins, and the detection is carried on each bin to determine whether there exists a significant frequency. In what follows, we define the detection and estimation problem related to the design of the RSFT.

Let SNRi=σbi2/σn2SNR_{i}=\sigma_{bi}^{2}/\sigma_{n}^{2} be the SNR of the ithi_{th} sinusoid. Let us define the worst case SNR, i.e., SNRminSNR_{min}, as

SNRminmini[K](SNRi).SNR_{min}\triangleq\mathop{\mathrm{min}}_{i\in[K]}(SNR_{i}). (12)

Let PdP_{d} denote for the probability of detection for the sinusoid with SNRminSNR_{min}, and PfaP_{fa} the corresponding probability of false alarm on each frequency bin.

Problem 1.

For the signal defined in (10), find the optimal thresholds of the first and the second stage of detection in an asymptotic sense, such that they minimize SNRminSNR_{min} for given Pd,Pfa,N,B,T,KP_{d},P_{fa},N,B,T,K. Also, characterize the tradeoff between computational complexity and SNRminSNR_{min} as a function of various parameters.

In the following, we investigate the two stages of detection separately, then summarize the solution into an optimization problem.

IV-B First Stage Detection

The first stage detection is performed on each data segment. After pre-permutation windowing, permutation and flat-windowing, the input signal can be expressed as

𝐳=𝐖¯𝐏σs𝐖𝐫,\mathbf{z}=\overline{\mathbf{W}}\mathbf{P}_{\sigma_{s}}\mathbf{W}\mathbf{r}, (13)

where σs\sigma_{s} is the permutation parameter for the sths_{th} segment, which has an uniform random distribution; 𝐏σs\mathbf{P}_{\sigma_{s}} is the permutation matrix, which functions as (2) with τ=0\tau=0; 𝐖=diag(𝐰)\mathbf{W}=\mathop{\mathrm{diag}}\nolimits(\mathbf{w}), 𝐖¯=diag(𝐰¯)\overline{\mathbf{W}}=\mathop{\mathrm{diag}}\nolimits(\bar{\mathbf{w}}), where 𝐰\mathbf{w} and 𝐰¯\bar{\mathbf{w}} are pre-permutation window and flat-window, respectively.

Regarding the design of the flat-window, we take its frequency domain main-lobe to have width 2π/B2\pi/B, and choose its length in time domain as NN. As indicated in [2], it is possible to use less data in flat-windowing by choosing the length of 𝐰¯\bar{\mathbf{w}} less than NN, i.e., dropping some samples in each segment after the permutation. However, a reduced length window in the time domain would result in longer transition regions in the frequency domain and as a result the detection performance of the system would degrade, since a larger transition region would allow more noise to enter the estimation process.

The time domain aliasing can be described as

𝐟=i[L](𝐖¯i𝐏σs𝐖𝐫)=𝐕σs𝐫,\mathbf{f}=\sum_{i\in[L]}\left(\overline{\mathbf{W}}_{i}\mathbf{P}_{\sigma_{s}}\mathbf{W}\mathbf{r}\right)=\mathbf{V}_{\sigma_{s}}\mathbf{r}, (14)

where L=N/BL=N/B; 𝐖¯i\overline{\mathbf{W}}_{i} is the ithi_{th} sub-matrix of 𝐖¯\overline{\mathbf{W}}, which is comprised of the iBthiB_{th} to the ((i+1)B1)th((i+1)B-1)_{th} rows of 𝐖¯\overline{\mathbf{W}}. And 𝐕σs=i[L]𝐖¯i𝐏σs𝐖\mathbf{V}_{\sigma_{s}}=\sum_{i\in[L]}\overline{\mathbf{W}}_{i}\mathbf{P}_{\sigma_{s}}\mathbf{W}.

The FFT operation on the aliased data 𝐟\mathbf{f} can be expressed as

𝐟^=𝐃𝐕σs𝐫,\hat{\mathbf{f}}=\mathbf{D}\mathbf{V}_{\sigma_{s}}\mathbf{r}, (15)

where 𝐃B×B\mathbf{D}\in\mathbb{C}^{B\times B} is the DFT matrix. For the kthk_{th} entry of 𝐟^\hat{\mathbf{f}}, we have

[𝐟^]k=𝐮kH𝐕σs𝐫,k[B],[\hat{\mathbf{f}}]_{k}=\mathbf{u}^{H}_{k}\mathbf{V}_{\sigma_{s}}\mathbf{r},\;k\in[B], (16)

where 𝐮k\mathbf{u}_{k} is the kthk_{th} column of 𝐃\mathbf{D}, i.e., 𝐮k=[1ejkΔωBejk(B1)ΔωB]T\mathbf{u}_{k}=[1\quad e^{jk\Delta\omega_{B}}\;\cdots\;e^{jk(B-1)\Delta\omega_{B}}]^{T}, and ΔωB=2π/B\Delta\omega_{B}=2\pi/B.

Substituting (10) into (16), and taking the mthm_{th} sinusoid, which we assume is the weakest sinusoid with its SNR equals to SNRminSNR_{min}, out of the summation

[𝐟^]k=bm𝐮kH𝐕σs𝐯(ωm)+j[K]m(bj𝐮kH𝐕σs𝐯(ωj))+𝐮kH𝐕σs𝐧.\begin{split}[\hat{\mathbf{f}}]_{k}&=b_{m}\mathbf{u}^{H}_{k}\mathbf{V}_{\sigma_{s}}\mathbf{v}(\omega_{m})\\ &+\sum_{j\in[K]\setminus m}\left(b_{j}\mathbf{u}^{H}_{k}\mathbf{V}_{\sigma_{s}}\mathbf{v}(\omega_{j})\right)\\ &+\mathbf{u}^{H}_{k}\mathbf{V}_{\sigma_{s}}\mathbf{n}.\end{split} (17)

Since [𝐟^]k[\hat{\mathbf{f}}]_{k} is a linear combination of bi,[𝐧]j,i[K],j[N]b_{i},[\mathbf{n}]_{j},i\in[K],j\in[N], it holds that

[𝐟^]k𝒞𝒩(0,σfk2),[\hat{\mathbf{f}}]_{k}\sim\mathcal{CN}(0,\sigma_{fk}^{2}), (18)

where

σfk2=σbm2α(k,σs,ωm)+j[K]m(σbj2α(k,σs,ωj))+σn2β(σs),\begin{split}\sigma_{fk}^{2}&=\sigma_{bm}^{2}\alpha(k,\sigma_{s},\omega_{m})\\ &+\sum_{j\in[K]\setminus m}\left(\sigma_{bj}^{2}\alpha(k,\sigma_{s},\omega_{j})\right)+\sigma_{n}^{2}\beta(\sigma_{s}),\end{split} (19)

and

α(k,σs,ω)=|𝐮kH𝐕σs𝐯(ω)|2β(σs)=𝐖¯𝐏σs𝐰2.\begin{split}&\alpha(k,\sigma_{s},\omega)=|\mathbf{u}^{H}_{k}\mathbf{V}_{\sigma_{s}}\mathbf{v}(\omega)|^{2}\\ &\beta(\sigma_{s})=\|\overline{\mathbf{W}}\mathbf{P}_{\sigma_{s}}\mathbf{w}\|^{2}.\end{split} (20)

It is easy to see that σfk2\sigma_{fk}^{2} is summation of weighted variance from each signal and noise component.

We now investigate the pthp_{th} bin, where ωm\omega_{m} is mapped to, and we have the following claim.

Claim 1.

For a complex sinusoid signal, i.e., 𝐯(ω)\mathbf{v}(\omega), after pre-permutation windowing, permutation with σs\sigma_{s}, flat windowing, aliasing and FFT, the highest amplitude of signal spectrum appears in [B][B] at location

p(ω,σs)=BN[σsωΔωN]N.p(\omega,\sigma_{s})=\lfloor{\frac{B}{N}[\sigma_{s}\lfloor\frac{\omega}{\Delta\omega_{N}}\rfloor]_{N}}\rfloor. (21)

where ΔωN=2π/N\Delta\omega_{N}=2\pi/N.

Proof.

If we were applying DFT to 𝐯(ω)\mathbf{v}(\omega), the highest amplitude of the spectrum would appear on the grid point closest to ω\omega, i.e. ωΔωN\lfloor\frac{\omega}{\Delta\omega_{N}}\rfloor. The pre-permutation windowing will not change the position of the highest peak, provided the window is symmetric. Then after permutation, the peak location dilates by σs\sigma_{s} modularly, and becomes [σsωΔωN]N[\sigma_{s}\lfloor\frac{\omega}{\Delta\omega_{N}}\rfloor]_{N}. Finally, after flat-windowing and aliasing, the signal is ideally downsampled in the frequency domain, and the data length changes from NN to BB. Then the BB-point DFT exhibits the highest peak on grid point pp as desired. A visualization of this process is shown in Fig. 4. ∎

Refer to caption
Figure 4: Windowing, permutation and aliasing. The frequency representation of the signal from pre-permutation windowing to aliasing is presented. Only one significant frequency is shown for conciseness.

On assuming that only ωm\omega_{m} maps to bin pp, and that the side-lobes (leakage) are far below the noise level (owning to the two stages of windowing, which attenuate the leakage down to a desired level), the effect of leakage from other sinusoids can be ignored. Then we can approximate the variance of [𝐟^]p[\hat{\mathbf{f}}]_{p} as

σfp2σbm2α(p,σs,ωm)+σn2β(σs).\begin{split}\sigma_{fp}^{2}&\approx\sigma_{bm}^{2}\alpha(p,\sigma_{s},\omega_{m})+\sigma_{n}^{2}\beta(\sigma_{s}).\end{split} (22)

In case that multiple frequencies are mapped to the same bin (collision), (22) gives a underestimate of the variance. The probability of a collision occurring reduces as K<<BK<<B.

The bin u[B]u\in[B], for which no significant frequency is mapped to, contains only noise, and the corresponding variance for [𝐟^]u[\hat{\mathbf{f}}]_{u} is

σfu2σn2β(σs).\begin{split}\sigma_{fu}^{2}&\approx\sigma_{n}^{2}\beta(\sigma_{s}).\end{split} (23)

Hence, the hypothesis test for the first-stage-detection on [𝐟^]j,j[B][\hat{\mathbf{f}}]_{j},j\in[B] is formulated as

  • 0\mathcal{H}0: no significant frequency is mapped to it.

  • 1\mathcal{H}1: at least one significant frequency is mapped to it, with worst case SNR equals to SNRminSNR_{min}.

The log likelihood ratio test (LLRT) is

logPfj|H1(x)Pfj|H0(x)01γ.\log{P_{f_{j}|H1}(x)\over P_{f_{j}|H0}(x)}\mathop{\gtrless}_{\mathcal{H}0}^{\mathcal{H}1}\gamma^{\prime}. (24)

where Pfj|H1(x)P_{f_{j}|H1}(x) and Pfj|H0(x)P_{f_{j}|H0}(x) are the probability density function (PDF) of [𝐟^]j[\hat{\mathbf{f}}]_{j} under 1\mathcal{H}1 and 0\mathcal{H}0 respectively, and γ\gamma^{\prime} is a threshold.

Substituting the PDF of [𝐟^]j[\hat{\mathbf{f}}]_{j} under both hypothesis into (24), and after some manipulations we get

|[𝐟^]j|201γlogσfu2σfp21σfu21σfp2.|[\hat{\mathbf{f}}]_{j}|^{2}\mathop{\gtrless}_{\mathcal{H}0}^{\mathcal{H}1}{{\gamma^{\prime}}-{{\log{\sigma^{2}_{f_{u}}\over\sigma^{2}_{f_{p}}}}}\over{{1\over\sigma^{2}_{f_{u}}}-{1\over\sigma^{2}_{f_{p}}}}}. (25)

Hence, |[𝐟^]j|2|[\hat{\mathbf{f}}]_{j}|^{2} is a sufficient statistics for first stage detection. Since [𝐟^]j[\hat{\mathbf{f}}]_{j} has circularly symmetric Gaussian distribution, |[𝐟^]j|2|[\hat{\mathbf{f}}]_{j}|^{2} is exponentially distributed with cumulative distribution function (CDF)

F|[𝐟^]j|2(x,ζ2)={1exζ2,x00,x<0,F_{|[\hat{\mathbf{f}}]_{j}|^{2}}(x,\zeta^{2})=\begin{cases}1-e^{-{x\over\zeta^{2}}},\;x\geq 0\\ 0,\;x<0\;,\end{cases} (26)

where ζ2\zeta^{2} equals to σfu2\sigma^{2}_{f_{u}} under 0{\mathcal{H}0} and σfp2\sigma^{2}_{f_{p}} under 1{\mathcal{H}1}.

Based on (26), in the first stage of detection, the false alarm rate on each of BB bins and the probability of detection of the weakest sinusoid can be derived to be equal to

P~fa(σs)=eγσn2β(σs),P~d(ωm,σs)=P~faβ(σs)α(p,ωm,σs)SNRmin+β(σs),\begin{split}&\tilde{P}_{fa}(\sigma_{s})=e^{-{\gamma\over\sigma^{2}_{n}\beta(\sigma_{s})}},\\ &\tilde{P}_{d}(\omega_{m},\sigma_{s})=\tilde{P}_{fa}^{\beta(\sigma_{s})\over\alpha(p,\omega_{m},\sigma_{s})SNR_{min}+\beta(\sigma_{s})},\end{split} (27)

where γ\gamma is the detection threshold. Both P~fa\tilde{P}_{fa} and P~d\tilde{P}_{d} depend on the permutation σs\sigma_{s}. Taking expectation with respect to σs\sigma_{s}, we have

P¯fa=eγσn2β¯,P¯d(ωm)=P¯faβ¯α¯(p,ωm)SNRmin+β¯,\begin{split}&\bar{P}_{fa}=e^{-{\gamma\over\sigma^{2}_{n}\bar{\beta}}},\\ &\bar{P}_{d}(\omega_{m})=\bar{P}_{fa}^{\bar{\beta}\over\bar{\alpha}(p,\omega_{m})SNR_{min}+\bar{\beta}},\end{split} (28)

where P¯d(ωm)=𝔼{P~d(σs,ωm)}\bar{P}_{d}(\omega_{m})=\mathbb{E}\{\tilde{P}_{d}(\sigma_{s},\omega_{m})\}, P¯fa=𝔼{P~fa(σs)}\bar{P}_{fa}=\mathbb{E}\{\tilde{P}_{fa}(\sigma_{s})\}, α¯(p,ωm)=𝔼{α~(p,ωm,σs)}\bar{\alpha}(p,\omega_{m})=\mathbb{E}\{\tilde{\alpha}(p,\omega_{m},\sigma_{s})\}, and β¯=𝔼{β~(σs)}\bar{\beta}=\mathbb{E}\{\tilde{\beta}(\sigma_{s})\}.

IV-C Second Stage Detection

Let 𝐜σs{0,1}B\mathbf{c}_{\sigma_{s}}\in\{0,1\}^{B} denote the output of the first-stage-detection for the sths_{th} segment, with permutation factor σs\sigma_{s}. Each entry in 𝐜σs\mathbf{c}_{\sigma_{s}} is a Bernoulli random variable, i.e., for j[B]j\in[B],

[𝐜σs]j{Bernoulli(P~fa(σs)),under0,Bernoulli(P~d(ωm,σs)),under1.{[\mathbf{c}_{\sigma_{s}}]_{j}}\sim\begin{cases}\mathrm{Bernoulli}\left(\tilde{P}_{fa}(\sigma_{s})\right),under\;\mathcal{H}0,\\ \mathrm{Bernoulli}\left(\tilde{P}_{d}(\omega_{m},\sigma_{s})\right),under\;\mathcal{H}1.\end{cases} (29)

Note that under 1\mathcal{H}1, we assume that [𝐜σs]j[\mathbf{c}_{\sigma_{s}}]_{j} corresponds to the weakest sinusoid. For the other K1K-1 co-existing sinusoids, since their SNR may be grater than SNRminSNR_{min}, their probability of detection may also be grater than P~d(ωm,σs)\tilde{P}_{d}(\omega_{m},\sigma_{s}) (see Claim 3).

The reverse-mapping stage hashes the BB-dimensional 𝐜σs\mathbf{c}_{\sigma_{s}} back to the NN-dimensional 𝐚σs\mathbf{a}_{\sigma_{s}}. According to Definition 4, it holds that

[𝐚σs]i=[𝐜σs]j,i[N],j[B],i(j,σs1).[\mathbf{a}_{\sigma_{s}}]_{i}=[\mathbf{c}_{\sigma_{s}}]_{j},\;i\in[N],j\in[B],i\in\mathcal{R}(j,\sigma_{s}^{-1}). (30)

After accumulation of TT iterations, each entry in the accumulated output is summation of TT Bernoulli variables with different success rate. Define 𝐚¯\bar{\mathbf{a}} as the accumulated output, then for its ith,i[N]i_{th},i\in[N] entry, we have

[𝐚¯]i=s[T][𝐚σs]i=i(j,σs1),s[T][𝐜σs]j.[\bar{\mathbf{a}}]_{i}=\sum_{s\in[T]}[\mathbf{a}_{\sigma_{s}}]_{i}=\sum_{i\in\mathcal{R}(j,\sigma_{s}^{-1}),s\in[T]}[\mathbf{c}_{\sigma_{s}}]_{j}. (31)

Note that in (31), each term inside the sum corresponds to a different segment, i.e., [𝐜σs]j[\mathbf{c}_{\sigma_{s}}]_{j} is from the sths_{th} segment. Since σs\sigma_{s} is drawn randomly for each segment, jj may take different values, and relates to ii via a reverse-mapping. Fig. 5 gives a graphical illustration of the mapping and reverse-mapping.

Now, the hypothesis test for the second-stage-detection on [𝐚¯]i,i[N][\bar{\mathbf{a}}]_{i},i\in[N] is formulated as

  • ¯0\overline{\mathcal{H}}0: no significant frequency exists.

  • ¯1\overline{\mathcal{H}}1: there exists a significant frequency, whose SNR is at least SNRminSNR_{min}.

In the following, we investigate the statistics of [𝐚¯]i[\bar{\mathbf{a}}]_{i} under both hypothesis in an asymptotic senses. Before that however, we will take a closer look at the mapping and the reverse mapping by providing the following properties.

Refer to caption
Figure 5: Mapping and Reverse Mapping. A significant frequency may map to different locations in different iterations of the first stage detection, due to the different permutations. The detected frequencies, including the false alarms in the first stage, are reverse mapped to the original dimension. The true location of the significant frequency is recovered, and the ambiguous frequencies are also generated. The occurrence on the true location grows steadily during accumulation, provided the SNR is high enough, and thus the true location can be recovered in the second stage of detection with proper thresholding. However, false alarms may also occur in the second stage detection, due to both ambiguous frequencies and false alarms from the first stage of detection.
Property 3.

(Reversibility): Let j[B],i,σ,σ1[N]j\in[B],i,\sigma,\sigma^{-1}\in[N]. σ\sigma and σ1\sigma^{-1} satisfy Eq. (1). If j=(i,σ)j=\mathcal{M}(i,\sigma), then it holds that

i(j,σ1).i\in\mathcal{R}(j,\sigma^{-1}). (32)
Property 4.

(Distinctiveness): Let i,j[B],iji,j\in[B],i\neq j. If σ1[N]\sigma^{-1}\in[N] and satisfies Eq. (1), then it holds that

(i,σ1)(j,σ1)=.\mathcal{R}(i,\sigma^{-1})\cap\mathcal{R}(j,\sigma^{-1})=\emptyset. (33)

The proofs of these properties are provided in Appendix B. The two properties simply reveal the following facts: a mapped location can be recovered by reverse mapping (with ambiguities). Also, when applying reverse mapping to two distinct locations with the same permutation parameter, the resulting locations are also distinct.

Under ¯1\overline{\mathcal{H}}1, assuming that [𝐚¯]i[\bar{\mathbf{a}}]_{i} corresponds to the mthm_{th} sinusoid, i.e., the weakest sinusoid, then each term inside the sum of (31) has distribution [𝐜σs]jBernoulli(P~d(ωm,σs)),s[T][\mathbf{c}_{\sigma_{s}}]_{j}\sim\mathrm{Bernoulli}\left(\tilde{P}_{d}(\omega_{m},\sigma_{s})\right),s\in[T]. Then we present the following claim.

Claim 2.

Under ¯1\overline{\mathcal{H}}1, and as TT\to\infty,

[𝐚¯]iN(μa1(ωm),σa12(ωm)),[\bar{\mathbf{a}}]_{i}\sim N(\mu_{a1}(\omega_{m}),\sigma^{2}_{a1}(\omega_{m})), (34)

where μa1(ωm)=TP¯d(ωm)\mu_{a1}(\omega_{m})=T\bar{P}_{d}(\omega_{m}), σa12(ωm)TP¯d(ωm)(1P¯d(ωm))\sigma^{2}_{a1}(\omega_{m})\leq T\bar{P}_{d}(\omega_{m})(1-\bar{P}_{d}(\omega_{m})).

Proof.

Since 0<P~d(ωm,σs)(1P~d(ωm,σs))<10<\tilde{P}_{d}(\omega_{m},\sigma_{s})(1-\tilde{P}_{d}(\omega_{m},\sigma_{s}))<1, for δ>0\delta>0, the Lyapunov Condition[18] holds, i.e.,

limT1σa1(ωm)2+δs[T]𝔼{|[𝐜σs]j(ωm)P~d(ωm,σs)|2+δ}limT1σa1(ωm)2+δs[T]𝔼{|[𝐜σs]j(ωm)P~d(ωm,σs)|2}=limT1σa1(ωm)δ=0.\begin{split}&\lim_{T\to\infty}\frac{1}{\sigma_{a1}(\omega_{m})^{2+\delta}}\sum_{s\in[T]}\mathbb{E}\{|[\mathbf{c}_{\sigma_{s}}]_{j}(\omega_{m})-\tilde{P}_{d}(\omega_{m},\sigma_{s})|^{2+\delta}\}\\ &\leq\lim_{T\to\infty}\frac{1}{\sigma_{a1}(\omega_{m})^{2+\delta}}\sum_{s\in[T]}\mathbb{E}\{|[\mathbf{c}_{\sigma_{s}}]_{j}(\omega_{m})-\tilde{P}_{d}(\omega_{m},\sigma_{s})|^{2}\}\\ &=\lim_{T\to\infty}\frac{1}{\sigma_{a1}(\omega_{m})^{\delta}}=0.\end{split} (35)

Therefore, [𝐚¯]i[\bar{\mathbf{a}}]_{i} conforms to the Normal distribution as indicated in (34). It also holds that

σa12(ωm)=s[T]P~d(ωm,σs)(1P~d(ωm,σs))=TP¯d(ωm)(1P¯d(ωm))s[T](P~d(ωm,σs)P¯d(ωm))2,\begin{split}\sigma^{2}_{a1}(\omega_{m})&=\sum_{s\in[T]}\tilde{P}_{d}(\omega_{m},\sigma_{s})(1-\tilde{P}_{d}(\omega_{m},\sigma_{s}))\\ &=T\bar{P}_{d}(\omega_{m})(1-\bar{P}_{d}(\omega_{m}))\\ &-\sum_{s\in[T]}(\tilde{P}_{d}(\omega_{m},\sigma_{s})-\bar{P}_{d}(\omega_{m}))^{2},\end{split} (36)

from which we get that σa12(ωm)TP¯d(ωm)(1P¯d(ωm))\sigma^{2}_{a1}(\omega_{m})\leq T\bar{P}_{d}(\omega_{m})(1-\bar{P}_{d}(\omega_{m})), with the equality holding when P~d(ωm,σs)=P¯d(ωm)\tilde{P}_{d}(\omega_{m},\sigma_{s})=\bar{P}_{d}(\omega_{m}).

The distribution of [𝐚¯]i[\bar{\mathbf{a}}]_{i} under ¯0\overline{\mathcal{H}}0 is more complicated, and we have following claim.

Claim 3.

Under ¯0\overline{\mathcal{H}}0, and as TT\to\infty,

[𝐚¯]iN(μa0(ωm),σa02(ωm)),[\bar{\mathbf{a}}]_{i}\sim N(\mu_{a0}(\omega_{m}),\sigma^{2}_{a0}(\omega_{m})), (37)

where

fμa0(ωm)=FηpP¯d(ωm)+(TF)P¯fa,σa02(ωm)FηpP¯d(ωm)(1ηpP¯d(ωm))+(TF)P¯fa(1P¯fa),\begin{split}f\mu_{a0}(\omega_{m})&=F\eta_{p}\bar{P}_{d}(\omega_{m})+(T-F)\bar{P}_{fa},\\ \sigma^{2}_{a0}(\omega_{m})&\leq F\eta_{p}\bar{P}_{d}(\omega_{m})(1-\eta_{p}\bar{P}_{d}(\omega_{m}))\\ &+(T-F)\bar{P}_{fa}(1-\bar{P}_{fa}),\end{split} (38)

and F=TKηmBF=\frac{TK\eta_{m}}{B}, where ηm\eta_{m} is the 6.0-dB bandwidth of the pre-permutation window 𝐰\mathbf{w}. ηp[1,1Pd¯(ωm)]\eta_{p}\in[1,\frac{1}{\bar{P_{d}}(\omega_{m})}] is a calibration factor of the probability of detection for the other K1K-1 co-existing sinusoids.

Proof.

Under ¯0\overline{\mathcal{H}}0, each term in (31) may be distributed differently. To illustrate this, we consider a location i[N]i\in[N] in the frequency domain of the input signal, which does not contain a significant frequency, as shown in Fig. 5. Let j=(i,σs)j=\mathcal{M}(i,\sigma_{s}) be the mapping. There would be two cases for jj: 1) jj does not contain a significant frequency; or 2) jj contains at least one significant frequency, with its SNR at least SNRminSNR_{min}. In the former case, [𝐜σs]jBernoulli(P~fa(σs)){[\mathbf{c}_{\sigma_{s}}]_{j}}\sim\mathrm{Bernoulli}\left(\tilde{P}_{fa}(\sigma_{s})\right), i.e., [𝐜σs]j{[\mathbf{c}_{\sigma_{s}}]_{j}} is under 0\mathcal{H}0. For the latter case, [𝐜σs]jBernoulli(P~d(ωm,σs)){[\mathbf{c}_{\sigma_{s}}]_{j}}\sim\mathrm{Bernoulli}\left(\tilde{P}_{d}(\omega_{m},\sigma_{s})\right), i.e., [𝐜σs]j{[\mathbf{c}_{\sigma_{s}}]_{j}} is under 1\mathcal{H}1. Due to the permutation being uniformly random, on the average, the number of [𝐜]j[\mathbf{c}]_{j} under 1\mathcal{H}1 is F=TKηmBF=\frac{TK\eta_{m}}{B}, and the number of [𝐜]j[\mathbf{c}]_{j} under 0\mathcal{H}0 is TFT-F. The parameter ηm\eta_{m} reflects the fact that sparsity is affected by the pre-permutation windowing. Since we assume that 𝐯(ωm)\mathbf{v}(\omega_{m}) has the minimum SNR, i.e., SNRminSNR_{min}, other sinusoids with higher SNR will have larger P¯d\bar{P}_{d}. Hence we multiply P¯d(ωm)\bar{P}_{d}(\omega_{m}) with ηp\eta_{p} to calibrate the successful rate of [𝐜σs]j[\mathbf{c}_{\sigma_{s}}]_{j} under 1\mathcal{H}1. If all the sinusoids’s SNR were equal to SNRminSNR_{min}, then ηp=1\eta_{p}=1; on the other hand, if the co-existing sinusoids’ SNR were sufficient high so that their P¯d\bar{P}_{d} approaches to 11, then ηp=1Pd¯(ωm)\eta_{p}=\frac{1}{\bar{P_{d}}(\omega_{m})}. Finally, the results follows immediately by applying Lyapunov CLT. ∎

Remark 1.

From Claim 2 and 3, we notice that for the second stage detection, the LLRT is obtained based on two Normal distributions. The test statistic under ¯1\overline{\mathcal{H}}1 is “stable”, for it only depends on P¯d(ωm)\bar{P}_{d}(\omega_{m}). However, under ¯0\overline{\mathcal{H}}0, the distribution depends on the number of co-existing sinusoids, as well as on each sinusoid’s SNR. The larger KK and higher SNR will “push” the distribution under ¯0\overline{\mathcal{H}}0 closer to the distribution under ¯1\overline{\mathcal{H}}1, hence degrades the detection performance. In order to compensate for this, a larger SNRminSNR_{min} is required.

A natural extension of Remark 1 is Remark 2, which gives the condition under which the RSFT will reach its limit.

Remark 2.

Assuming that PdPfaP_{d}\geq P_{fa}, the RSFT will fail if KηmBK\eta_{m}\geq B no matter how large the SNRminSNR_{min} is.

Proof.

Assuming ηp=1\eta_{p}=1 and substituting Kηm=BK\eta_{m}=B into FF yields F=TF=T, which means that the distributions under both hypothes are the same, hence the two hypothesis cannot be discriminated. If ηp>1\eta_{p}>1, the assumption of PdPfaP_{d}\geq P_{fa} will be violated as KηmK\eta_{m} approaching BB. ∎

Based on the above discussion, the optimal threshold design in Problem 1 can be solved by the following optimization problem, i.e.,

Minimize{μ,P¯fa,P¯d}SNRminSubjecttoP¯d(ωm)=P¯faβ¯α¯(p,ωm)SNRmin+β¯Pfa=μga0(u)𝑑uPd=μga1(u)𝑑u0P¯fa1, 0P¯d1μ[T],\begin{split}&Minimize_{\{\mu,\bar{P}_{fa},\bar{P}_{d}\}}\quad SNR_{min}\\ &Subject\;to\\ &\quad\quad\bar{P}_{d}(\omega_{m})=\bar{P}_{fa}^{\bar{\beta}\over\bar{\alpha}(p,\omega_{m})SNR_{min}+\bar{\beta}}\\ &\quad\quad P_{fa}=\int_{\mu}^{\infty}g_{a_{0}}(u)du\\ &\quad\quad P_{d}=\int_{\mu}^{\infty}g_{a_{1}}(u)du\\ &\quad\quad 0\leq\bar{P}_{fa}\leq 1,\;0\leq\bar{P}_{d}\leq 1\\ &\quad\quad\mu\in[T],\end{split} (39)

where ga0(u),ga1(u)g_{a_{0}}(u),g_{a_{1}}(u) are the asymptotic PDF222We take the upper bounds of the variances in both distributions. It is shown in Section V-E that the actual variances is close to their upper bounds. of [𝐚¯]i[\bar{\mathbf{a}}]_{i} (which corresponds the weakest sinusoid) under ¯0\overline{\mathcal{H}}0 and ¯1\overline{\mathcal{H}}1, respectively. Since both of them are Normal distributions, with fixed threshold, i.e., μ\mu, we can solve for P¯d(ωm),P¯fa\bar{P}_{d}(\omega_{m}),\bar{P}_{fa}, and then compute the SNRminSNR_{min}. By enumerating μ[T]\mu\in[T], the minimum worst case SNR, i.e., SNRminSNR_{min}^{\ast} can be found, and the corresponding μ\mu^{\ast} is the optimal threshold for the second stage of detection. The optimal threshold for the first stage of detection, i.e., γ\gamma^{\ast}, can thus be calculated via (28).

Remark 3.

In Claim 3, we set a parameter ηp\eta_{p} to calibrate the distribution of [𝐚¯]i[\bar{\mathbf{a}}]_{i} under ¯0\overline{\mathcal{H}}0. By setting ηp\eta_{p} as 11 or 1Pd¯(ωm)\frac{1}{\bar{P_{d}}(\omega_{m})}, we can get respectively the lower and upper bound of SNRminSNR_{min}^{\ast} for the variation of SNR of other co-existing sinusoids. If KK is the maximum budget of signal sparsity, the optimal thresholds found by solving (39) provides the optimal thresholds for the worst case. If the actual signal sparsity were less than KK, PfaP_{fa} would be lower than the expected value, while PdP_{d} would be unchanged according to Remark 1.

By averaging over the permutation, asymptotically, SNRminSNR_{min}^{\ast} does not depend on the permutation. However, it still depends on ωm\omega_{m}, and we have the following claim to manifest their relationship.

Claim 4.

The dependence of SNRminSNR_{min}^{\ast} on ωm\omega_{m} is due to the off-grid loss[16] from off-grid frequencies. SNRminSNR_{min}^{\ast} attains its minimum when ωm\omega_{m} is on-grid, i.e. ωm=kΔωN,k[N]\omega_{m}=k\Delta\omega_{N},k\in[N]. When ωm\omega_{m} is in the middle of two grid points, i.e., ωm=(k+12)ΔωN\omega_{m}=(k+\frac{1}{2})\Delta\omega_{N}, SNRminSNR_{min}^{\ast} attains its maximum.

Proof.

Assume 𝐫=𝐯(ωm)\mathbf{r}=\mathbf{v}(\omega_{m}). Since the pre-permutation window 𝐰\mathbf{w} is symmetric, if we applied DFT to the pre-permuted data, the amplitude of the spectrum would attain its maximum and minimum respectively when ωm\omega_{m} is on-grid or in the middle between two grid points. The subsequent permutation operation would not change the amplitude of the spectrum. Also, since the flat-window is used, the downsampling in the frequency domain, which is a result of aliasing, will not affect the amplitude either. The on-grid frequency generates highest amplitude, while the frequency in the middle of between grid points has the lowest amplitude. As a result, the two detection stages require the lowest SNR for on-grid frequencies, and the highest SNR for frequencies lying in the middle of between grid points. ∎

IV-D Tradeoff between Worst Case SNR and Complexity

IV-D1 Comparison to Bartlett Method

We compare the complexity of the RSFT with the FFT-based Bartlett method (see Appendix A) by counting the number of operations in both algorithms as shown in Table (I) and Table (II). The RSFT has complexity equal to

𝒪(T(N+B+BlogB+KηmNBηp)+N),\mathcal{O}\left(T(N+B+B\log B+\frac{K\eta_{m}N}{B\eta_{p}})+N\right), (40)

while the Bartlett method has complexity equal to 𝒪(TN(1+logN)+N)\mathcal{O}\left(TN(1+\log N)+N\right). Fig. 6 compares the RSFT’s complexity to that of Bartlett’s for various BB and KK. One can see that the RSFT enabled savings are remarkable when BB is chosen properly. Specifically, from Fig. 6 one can see, the lowest complexity for KK equals to 5,50,1005,50,100 is achieved when BB equals to 32,64,12832,64,128, respectively. Note that the core operation in RSFT is still FFT-based, but on a reduced dimension space. By leveraging the existing high performance FFT libraries such as FFTW [19], the implementation of the RSFT algorithm could be further improved.

Remark 4.

The complexity of RSFT is linearly depend on N,T,K,1/ηpN,T,K,1/\eta_{p} and ηm\eta_{m}, hence it is beneficial to choose a pre-permutation window with a small ηm\eta_{m}, provided the attenuation of the side-lobes is sufficient. We can also choose the optimal BB from (40) to minimize the computation. However, there are two additional constrains for BB, one is BB should be a power of 2, the other is KηmBK\eta_{m}\geq B, as stated in Remark 2.

TABLE I: Computational Complexity of Bartlett Method
Procedure Number of Operations
Windowing TNTN
FFT TN2logNT\frac{N}{2}\log N
Square TNTN
Detection NN
Complexity 𝒪(TN(1+logN)+N)\mathcal{O}\left(TN(1+\log N)+N\right)
TABLE II: Computational Complexity of RSFT
Procedure Number of Operations
Pre-Permutation Win TNTN
Permutation TNTN
Flat-Win TNTN
Aliasing TB(N/B1)TB(N/B-1)
FFT TB2logBT\frac{B}{2}\log B
Square TBTB
First-Stage-Detection TBTB
Reverse-Mapping TKηmNBηp\frac{TK\eta_{m}N}{B\eta_{p}}
Second-stage-Detection NN
Complexity 𝒪(T(N+B+BlogB+KηmNBηp)+N)\mathcal{O}\left(T(N+B+B\log B+\frac{K\eta_{m}N}{B\eta_{p}})+N\right)
Refer to caption
Figure 6: Comparison of Complexity. N=1024,T=50,ηm=1.4,ηp=1,B={8,16,32,64,128,256,512,1024}N=1024,T=50,\eta_{m}=1.4,\eta_{p}=1,B=\{8,16,32,64,128,256,512,1024\}.

IV-D2 Worst Case SNR and Complexity Trade Off

The reduced complexity of RSFT is achieved at a the cost of an increased SNRminSNR_{min}, which decreases the ability of detecting weak signals. The tradeoff between SNRminSNR_{min} and complexity for various choices of parameters is shown in Fig. 7. The performance of the Bartlett method is also shown as a reference. From Fig. 7 we can see that BB plays a central role in trading off SNRminSNR_{min} and complexity. A proper choice of BB can enhance the computational efficiency significantly with a reasonable increase of SNRminSNR_{min}. Also, since the sparsity KK affects both SNRminSNR_{min} and complexity, a less sparse signal will worsen both. The complexity of RSFT is larger than that of the Bartlett method by setting B=NB=N, due to the additional processing in the algorithm. Also, it cannot achieve the same SNRminSNR_{min} as the Bartlett method does.

Refer to caption
Figure 7: Worst Case SNR and Complexity Trade off. N=1024,T=50,ηm=1.4,ηp=1,B={8,16,32,64,128,256,512,1024},Pd=0.9,Pfa=106,K={5,10,100},ωm=ΔωN/2N=1024,T=50,\eta_{m}=1.4,\eta_{p}=1,B=\{8,16,32,64,128,256,512,1024\},P_{d}=0.9,P_{fa}=10^{-6},K=\{5,10,100\},\omega_{m}=\Delta\omega_{N}/2. The red dot shows the performance of the Bartlett method, which serves as a reference.

V Numerical Results

In this section, we verify our theoretical findings via simulations. We use the following common parameters for various settings, unless we state specifically. We take the following values: N=1024,T=50,B=64,ηm=1.8,Pd=0.9,Pfa=106,ωm=64.5ΔωN0.4N=1024,T=50,B=64,\eta_{m}=1.8,P_{d}=0.9,P_{fa}=10^{-6},\omega_{m}=64.5\Delta\omega_{N}\approx 0.4. We use a Dolph-Chebyshev window with 40dB40dB attenuation as pre-permutation windowing. The flat-window is also based on this window, and we set its passband width as 1/B1/B.

V-A Lower and Upper Bounds of SNRminSNR_{min}^{\ast} for Fixed Sparsity

According to Remark 3, we can calculate the lower bound and the upper bound of SNRminSNR_{min}^{\ast} and their corresponding thresholds for fixed sparsity. Fig. 8 and 9 shows the thresholding of RSFT with both bounds. We mark the amplitude of ωm\omega_{m} with a magenta dot in each figure.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 8: Lower Bound of Optimal Thresholds for Fixed Sparsity. K=4,SNRmin10.1dB,γ3.0×103,μ/T0.3K=4,SNR_{min}^{\ast}\approx-10.1dB,\gamma^{\ast}\approx 3.0\times 10^{-3},\mu^{\ast}/T\approx 0.3 (a) Spectrum from Bartlett method. (b) First-stage-detection. (c) Second-stage-detection. Data and threshold is normalized by TT. (d) Final output of RSFT.
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 9: Upper Bound of Optimal Thresholds for Fixed Sparsity. K=4,SNRmin9.1dB,γ2.5×103,μ/T0.4K=4,SNR_{min}^{\ast}\approx-9.1dB,\gamma^{\ast}\approx 2.5\times 10^{-3},\mu^{\ast}/T\approx 0.4 The co-existing sinusoids’s SNR is 10dB10dB higher than SNRminSNR_{min} (a) Spectrum from Bartlett method. (b) First-stage-detection. (c) Second-stage-detection. Data and threshold is normalized by TT. (d) Final output of RSFT.

V-B Unknown Signal Sparsity

Since we do not assume that we know the exact sparsity of the signal, we will use a guess for KK. Fig. 10 shows the optimal design was toward K=10K=10, however, when the true sparsity is K=3K=3, the system yields the same PdP_{d} but better PfaP_{fa}, since the noise level is much lower than expected.

Refer to caption
(a)
Refer to caption
(b)
Figure 10: Unknown Sparsity. The threshold is optimal for K=10K=10 (a) Second-stage-detection when K=10K=10. (b) Second-stage-detection when K=3K=3 with the same threshold.

V-C Dependency on Frequency

Fig. 11 shows the dependency of SNRminSNR_{min}^{\ast} on frequency, which verifies Claim 4.

Refer to caption
Figure 11: Dependence of SNRminSNR_{min}^{\ast} on Frequency. The fluctuation of SNRminSNR_{min}^{\ast} across the spectrum is mainly due to the off-grid loss of off-grid frequencies.

V-D The Receiver Operating Characteristic (ROC) Curve

In this section, we use ROC curve to characterize the performance of RSFT with variance parameters. Fig. 12 shows the impact of the detection by adopting different values of BB. A smaller BB lowers the detection performance, and in order to compensate it, a higher SNRminSNR_{min} is required. The ROC curve for the Bartlett method is calculated by (50) and is also shown in Fig. 12.

Fig. 13 illustrates the relationship between detection performance and sparsity of signal. It is shown that with other parameters fixed, the sparser the signal is, the better the performance of detection is.

In Fig. 14, we can see the impact of the SNR from the co-existing sinusoids. The higher the SNR of the co-existing sinusoids is, the worse the detection performance is.

Refer to caption
Figure 12: ROC Corresponding to Different Values of BB and SNRminSNR_{min}. ηp=1,K=4\eta_{p}=1,K=4.
Refer to caption
Figure 13: ROC Corresponding to Different Values of KK. ηp=1,B=256,SNRmin=20dB\eta_{p}=1,B=256,SNR_{min}=-20dB.
Refer to caption
Figure 14: ROC Corresponding to Different Values of SNR for Co-Existing Sinusoids. ηp={1,1/P¯d(ωm)},B=256,K=50,SNRmin=20dB\eta_{p}=\{1,1/\bar{P}_{d}(\omega_{m})\},B=256,K=50,SNR_{min}=-20dB.

V-E The Variance and Its Upper Bound for [𝐚¯]i[\bar{\mathbf{a}}]_{i}

In solving (39), we take the upper bound of the variance for [𝐚¯]i[\bar{\mathbf{a}}]_{i} under both hypotheses. In this section, we show by simulation that the actual variance of [𝐚¯]i[\bar{\mathbf{a}}]_{i} is close to its upper bound. In what follows, we study σa12(ωm)\sigma_{a1}^{2}(\omega_{m}); the case for σa02(ωm)\sigma_{a0}^{2}(\omega_{m}) can be similarly studied.

As shown in (36), the discrepancy of σa12(ωm)\sigma_{a1}^{2}(\omega_{m}) from its upper bound is due to the P~d\tilde{P}_{d}’s dependence on σs\sigma_{s}, which is caused by β\beta and α\alpha’s dependence on σs\sigma_{s} (see (27)). For NN a power of 2, a valid σs\sigma_{s} can be any odd integer in [N][N][2]. Fig. 15 shows β(σs)\beta(\sigma_{s}) and α(p,ωm,σs)\alpha(p,\omega_{m},\sigma_{s}) as functions of σs\sigma_{s}. The symmetry of the plot is due to the symmetry of pre-permutation window and the flat-window, as well as the modulo property of the permutation. Another observation is that most of β(σs)\beta(\sigma_{s}) and α(p,ωm,σs)\alpha(p,\omega_{m},\sigma_{s}) have similar values. As a result, P~d(ωm,σs)\tilde{P}_{d}(\omega_{m},\sigma_{s}) has similar value for different permutations, and this is the reason for σa12(ωm)\sigma_{a1}^{2}(\omega_{m}) being close to its upper bound. The Monte Carlo simulation in Fig. 16 shows that the approximation error, i.e., TP¯d(ωm)(1P¯d(ωm))σa12(ωm)σa12(ωm)\frac{T\bar{P}_{d}(\omega_{m})(1-\bar{P}_{d}({\omega_{m}}))-\sigma_{a1}^{2}(\omega_{m})}{\sigma_{a1}^{2}(\omega_{m})} decreases as TT grows, and even for a small TT, such as T=10T=10, the error is as small as about 1.6%1.6\%.

Refer to caption
Figure 15: β\beta and α(ωm)\alpha(\omega_{m})’s dependence on permutation. B=64B=64.
Refer to caption
Figure 16: Approximation Error for the Variance. B=64,T={10,50,100,200}B=64,T=\{10,50,100,200\}. Each value is averaged over 100 times Monte Carlo simulation.

VI RSFT for Ubiquitous Radar Signal Processing

The RSFT algorithm can greatly reduce the complexity of certain high dimensional problems. This can be signifiant in many applications, since lower complexity means faster reaction time and more economical hardware. However, in order to apply RSFT, the signal to be processed should meet the following requirements:

  • It should be sparse in some domain.

  • It should be sampled uniformly whether in temporal or spacial domain.

  • The SNR should be moderately high.

While many applications satisfy these requirements, in what follows, we discuss an example in Short Range Ubiquitous Radar[20] (SRUR) signal processing.

VI-A Short Range Ubiquitous Radar

An ubiquitous radar or SIMO radar can see targets everywhere at anytime without steering its beams as a traditional phased array radar does. In SRUR, a broad transmitting beam patten is achieved by an omnidirectional transmitter and multiple narrow beams are formed simultaneously after receiving of the reflected signal. The beam pattens of an ubiquitous radar is shown in Fig. 17 with an Uniform Linear Array (ULA) configuration.

Refer to caption
Figure 17: Ubiquitous Radar System Structure and Transmitting / Receiving Beam Patten. A broad beam patten is transmitted with an omnidirectional antenna, while multiple narrow beams are formed simultaneously by the receiving array. Each receiving channel is mixed with a coupled signal from the transmitter to de-chirp the LFMCW signal, before the A/D converting.

An SRUR with range coverage of several kilometers could be important both in military and civilian vehicular applications. For instance, in an active protection system [21], sensors on the protected vehicle have to detect and locate the warheads from a closely fired rocket-propelled grenade (RPG) within milliseconds. Among other sensors, SRUR’s simultaneous wide angle coverage, high precision of measurement and all-weather operation make it the ideal sensor for such situation.

Refer to caption
Figure 18: LFMCW Waveform. The signal frequency change linearly in time with a repetition internal TpT_{p}. A burst contains MM repetitions. The range of frequency changing is the bandwidth of the system. The received signal is a delayed version of the transmitted signal.

In order to achieve high range resolution and cover near range, SRUR utilizes a LFMCW waveform, as shown in Fig. 18. Mathematically, the transmitted waveform can be expressed as

s(t,v)=Acos(2π(fc(tvTp)+πρ(tvTp)2),s(t,v)=A\cos(2\pi(f_{c}(t-vT_{p})+\pi\rho(t-vT_{p})^{2}), (41)

where TpT_{p} is the repetition interval (RI), v[M]v\in[M] denotes the vthv_{th} RI, AA is amplitude of the signal, fcf_{c} is the carrier frequency and ρ\rho is the chirp rate. Furthermore, without loss of generality, we assume that the initial phase of the signal is zero.

Upon reception, a de-chirp process is implemented by mixing the received signal with the transmitted signal, followed by a lowpass filter. The received signal is a delayed version of the transmitted one, hence by mixing the two signals, the range information of the targets is linearly encoded in the difference of the frequencies. Hence for the ith,i[N]i_{th},i\in[N] receiving channel, the de-chirped signal is expressed as

ri=k[K]a[k](s)cos(2π((fr[k]+fd[k])(tvTp)+iπsinθ[k])+n(t),\begin{split}r_{i}&=\sum_{k\in[K]}a^{[k]}(s)\cos\left(2\pi((f_{r}^{[k]}+f_{d}^{[k]})(t-vT_{p})+i\pi\sin\theta^{[k]}\right)\\ &+n(t),\end{split} (42)

which is a superposition of KK sinusoids and additive noise n(t)n(t). For the kthk_{th} sinusoid, a[k](s),s[T]a^{[k]}(s),s\in[T] represents its amplitude, which can be modeled as a Gaussian random process. More specifically, the amplitude is assumed static within a burst, and independent between each of TT bursts. This assumption is consistent with the Swerling-I target model[22], which represents a slow fluctuation of the target RCS. fr[k],fd[k]f_{r}^{[k]},f_{d}^{[k]} are the frequency components respect to target’s range and velocity respectively, i.e.,

fr[k]=2ρrt[k]c,fd[k]=2vt[k]λ,f_{r}^{[k]}=\frac{2\rho r_{t}^{[k]}}{c},\quad f_{d}^{[k]}=\frac{2v_{t}^{[k]}}{\lambda}, (43)

where rt[k],vt[k],cr_{t}^{[k]},v_{t}^{[k]},c are the kthk_{th} target’s range, velocity and speed of wave propagation respectively.

The DOA of the kthk_{th} target, i.e., θ[k]\theta^{[k]} is defined as the angle between the line of sight (from the array center to the target) and the array normal. Assuming that the element wise spacing is λ/2\lambda/2, under the narrowband signal assumption, θ[k]\theta^{[k]} will cause an increase of phase at the neighboring array element equal to πsinθ[k]\pi\sin\theta^{[k]}. We omit the constant phase term in each sinusoids of (42), since they are irrelevant to the performance of the algorithm.

After AD conversion of each receiving channel, we can use the processing scheme shown in Fig. 19 to detect the targets as well as estimate their range, velocity and DOA. More specifically, grid-based versions of fr[k],fd[k],πsinθ[k]f_{r}^{[k]},f_{d}^{[k]},\pi\sin\theta^{[k]} can be calculated by applying a 3-D FFT on the windowed data cube, then, after accumulation of TT iterations, the above described NP detection procedure can be performed.

VI-B RSFT-based SRUR Signal Processing

Although the number of samples of SRUR is reduced significantly with the analog de-chirp processing, the realtime processing with 3-D FFT is still challenging. The RSFT algorithm is suitable for reducing the computational complexity of SRUR, since, 1) the number of targets is usually much smaller than the number of spatial resolutions cells, which implies that the signal is sparse after proper translation; 2) with a ULA and digitization of each received element, the signal is uniformly sampled both in spatial and temporal domain; and 3) the short range coverage implies that moderate high SNR is easy to achieve.

The RSFT-based SRUR processing architecture is shown in Fig. 20. Compared to the conventional processing, the 3-D FFT is replaced with a 3-D RSFT, in which the aliasing procedure converts the data cube dimensions from R×N×MR\times N\times M to B×C×DB\times C\times D. The 3-D FFT operated on the smaller data cube could save the computation time significantly.

Refer to caption
Figure 19: Conventional FFT-based Processing Scheme for SRUR.
Refer to caption
Figure 20: RSFT Based Processing Scheme for SRUR.

VI-C Simulations

In this section, we verify the feasibility of RSFT-based SRUR processing and compare to the SFT-based processing via simulations. The main parameters of the system are listed in Table III. The design of the system can guarantee non-ambiguous measurements of the target’s range and velocity, assuming the maximum range and velocity are less than 1.5km1.5km and 300m/s300m/s, respectively.

TABLE III: SRUR Parameters
Parameter Symbol Value
Number of range bins RR 20482048
Number of receiving elements NN 6464
Number of RI MM 3232
Wave length λ\lambda 0.03m0.03m
Wave propagation speed cc 3×108m/s3\times 10^{8}m/s
Bandwidth BwB_{w} 150MHz150MHz
Repetition interval TpT_{p} 5×105s5\times 10^{-5}s
Maxima range RmaxR_{max} 1.5×103m1.5\times 10^{3}m
Chirp rate ρ\rho 3×1012Hz/s3\times 10^{12}Hz/s
Sampling frequency (IQ) fsf_{s} 41MHz41MHz

We generate signals from 44 targets according to (42). The range, velocity and DOA of targets can be arbitrarily chosen within the unambiguous space, which implies that the corresponding frequency components do not necessarily lie on the grid. The targets’ parameters used in the simulation are listed in Table IV. For Targets 33 and 44, we set the same parameters except that their DOA is 44^{\circ} apart, which is close to the theoretical angular resolution after windowing for the Bartlett beamforming. To compare the RSFT and the SFT for different scenarios, we adopt two sets of SNR for targets. Specifically, for the first set, we use the same SNR, i.e., 10dB-10dB for different targets. And for the second set, we assign different SNR for different targets, which is more close to a realistic scenario.

TABLE IV: Target Parameters
Target Range (mm) Velocity (m/sm/s) DOA (\circ) SNR (dB)
1 10001000 100100 3030 10/0-10/0
2 500500 5050 0 10/10-10/-10
3 350350 240240 16-16 10/20-10/-20
4 350350 240240 20-20 10/20-10/-20

The SFT from [2] is 11-dimensional. In order to reconstruct targets in the 3-D space, we extend the SFT to high dimension with the techniques described in Section III-C. Another obstacle of applying SFT is that it needs to know the number of peaks to count in the detection stages. In our radar example, even the knowledge of exact number of targets is presented, it is still not clear how to determine the number of counting peaks due to existence of the large number of peaks from leakage. In the experiment, we gradually increase the number of counting peaks until all the targets are recovered. For the case of the same SNR setting, all the targets are recovered after around 2020 peaks are counted. While for the second SNR setting, we need to count nearly 200200 peaks to recover the weakest targets (Targets 33 and 44). Fig. 21 and 22 show the targets reconstruction results for the two settings, respectively. The former shows both SFT and RSFT methods can perfectly recover all the targets, whose SNR has the same value. From Targets 33 and 44 we can see that the SFT-based method achieves a better resolution than its RSFT counterpart, since the former does not require a pre-permutation window. For the second scenario, the SFT-based method shows the side-lobes of the stronger targets, while the RSFT-based method only recovers the (extended) main-lobes of all the targets.

The simulation shows that the RSFT-based approach is better than its SFT counterpart for a realistic scenario, within which the signal has a reasonable dynamic range. We also want to emphasis that in a real radar system, determine the number of counting peaks for the SFT-based method lacks a theoretical foundation, while the thresholding approach in the RSFT is consistent with the conventional FFT-based processing, both of which are based on the NP criterion.

Refer to caption
Figure 21: Target Reconstruction via 3-D SFT and RSFT with The Same SNR for 4 Targets. Both SFT and RSFT based methods can reconstruct all the targets. The former has better resolution without pre-permutation windowing.
Refer to caption
Figure 22: Target Reconstruction via 3-D SFT and RSFT with Different SNR for 4 Targets. The SFT-based processing recovers the side-lobes of the stronger targets, while the RSFT-based method only recovers the main-lobes of targets.

VII Conclusion

In this paper, we have addressed practical problems of applying SFT in real-life applications, based on that, we have proposed a modified algorithm (RSFT). The optimal design of parameters in RSFT has been analyzed, and the relationship between system sensitivity and computational complexity has been investigated. Some interesting properties of the RSFT have also been revealed by our analysis, such as the performance of detection not only relies on the frequency under examination, but also depends on other co-existing significant frequencies, which is very different from the traditional FFT-based processing. The analysis has revealed that RSFT could provide engineers an extra freedom of design in trading off system’s ability of detecting weak signals and complexity. Finally, the context of the application of RSFT has been discussed, and a specific example for short range ubiquitous radar signal processing has been presented.

Appendix A Bartlett Method Analysis

Detecting each sinusoid and estimating the corresponding frequencies in (10) can be achieved by Bartlett spectrum analysis followed by an NP detection. A window is applied on each data segment to reduce the leakage of off-grid frequencies. In order to enhance the computational efficiency, the FFT is adopted (see Algorithm 2). In the NN-dimensional case, each step in Algorithm 2 is done in NN-dimension.

The analysis of the Bartlett method can be found in the literature [23], however, such analysis usually focuses on bias, variance and frequency resolution, while the detection performance has not been throughly studied in connection with an NP detector. In [24], the performance of detecting a single sinusoid is discussed and the theoretical analysis is provided for on-grid frequency. However, most typically, the signal contains multiple significant frequencies, which are off-grid. In what follows, we analyze the asymptotic performance of Algorithm 2 based on the signal model of (10), which is a multiple-frequencies case and does not restrict the frequency being on-grid. Moreover, as compared to the signal model in [24], which assumes that the sinusoid has a deterministic amplitude, we model the complex amplitude of each sinusoid as a circularly symmetric Gaussian random variable. This modeling reflects the stochastic nature of each sinusoid, and is consistent with the Swerling-I target model in radar signal cases, since the square of a circularly symmetric Gaussian random variable has an exponential distribution.

Algorithm 2 Bartlett Method algorithm

Input: complex signal 𝐫\mathbf{r} in any fixed dimension
Output: 𝐨\mathbf{o}, frequency domain representation of input signal

1:procedure Bartlett(𝐫\mathbf{r})
2:  𝐱0\mathbf{x}\leftarrow 0
3:  for i=0Ti=0\to T do
4:   Windowing: 𝐮𝐖𝐫\mathbf{u}\leftarrow\mathbf{W}\mathbf{r}
5:   N-D FFT: 𝐮^\hat{\mathbf{u}} \leftarrow NDFFT(𝐮)\mathop{\mathrm{NDFFT}}(\mathbf{u})
6:   Accumulation: 𝐱𝐱+|𝐮^|2\mathbf{x}\leftarrow\mathbf{x}+|\hat{\mathbf{u}}|^{2}
7:  end for
8:  Detection: 𝐨NPdet(𝐱)\mathbf{o}\leftarrow\mathop{\mathrm{NPdet}}(\mathbf{x})
9:  return 𝐨\mathbf{o}
10:end procedure

The analysis of Algorithm 2 follows a similar fashion with the analysis of the RSFT, and we thus use the same notation as in Section IV. Our goal is to derive the relationship between PdP_{d} and PfaP_{fa}, which is also related to the worst case signal SNR, i.e., SNRminSNR_{min}.

After windowing and FFT, the signal becomes

𝐮^s=𝐃¯𝐖𝐫s,s[T],\hat{\mathbf{u}}_{s}=\overline{\mathbf{D}}\mathbf{W}\mathbf{r}_{s},\;s\in[T], (44)

where 𝐃¯N×N\overline{\mathbf{D}}\in\mathbb{C}^{N\times N} is the DFT matrix.

Substituting (10) into (44), for the kth,k[N]k_{th},k\in[N] entry of 𝐮^s\hat{\mathbf{u}}_{s}, we get

[𝐮^s]k=bm,s𝐝kH𝐖𝐯(ωm)+j[K]m(bj,s𝐝kH𝐖𝐯(ωj))+𝐝kH𝐖𝐧s,\begin{split}[\hat{\mathbf{u}}_{s}]_{k}&=b_{m,s}\mathbf{d}^{H}_{k}\mathbf{W}\mathbf{v}(\omega_{m})\\ &+\sum_{j\in[K]\setminus m}(b_{j,s}\mathbf{d}^{H}_{k}\mathbf{W}\mathbf{v}(\omega_{j}))\\ &+\mathbf{d}^{H}_{k}\mathbf{W}\mathbf{n}_{s},\end{split} (45)

where 𝐝k\mathbf{d}_{k} is the kthk_{th} column of 𝐃¯\overline{\mathbf{D}}, i.e., 𝐝k=[1ejkΔωNejk(N1)ΔωN]T\mathbf{d}_{k}=[1\quad e^{jk\Delta\omega_{N}}\;\cdots\;e^{jk(N-1)\Delta\omega_{N}}]^{T}, and ΔωN=2π/N\Delta\omega_{N}=2\pi/N.

Since [𝐮^s]k[\hat{\mathbf{u}}_{s}]_{k} is a linear combination of bi,s,[𝐧s]j,i[K],j[N]b_{i,s},[\mathbf{n}_{s}]_{j},i\in[K],j\in[N], it is a circularly symmetry Gaussian scalar with distribution

[𝐮^s]k𝒞𝒩(0,σuk2).[\hat{\mathbf{u}}_{s}]_{k}\sim\mathcal{CN}(0,\sigma_{uk}^{2}). (46)

The hypothesis test on each frequency bin is formulated as

  • 0\mathcal{H}^{\prime}0: no significant frequency exists.

  • 1\mathcal{H}^{\prime}1: there exists a significant frequency, with its SNR at least equals to SNRminSNR_{min}.

We assume the side-lobes of the significant frequencies are far below the noise level due to windowing, then under 1\mathcal{H}^{\prime}1 and 0\mathcal{H}^{\prime}0, respectively, we have the following approximation for σuk2\sigma_{uk}^{2}

σuk|H12σbm2α,σfk|H02σn2β.\begin{split}&\sigma_{uk|H1}^{2}\approx\sigma_{bm}^{2}\alpha^{\prime},\\ &\sigma_{fk|H0}^{2}\approx\sigma_{n}^{2}\beta^{\prime}.\\ \end{split} (47)

where α=|𝐝kH𝐖𝐯(ωm)|2\alpha^{\prime}=|\mathbf{d}^{H}_{k}\mathbf{W}\mathbf{v}(\omega_{m})|^{2} and β=𝐰2\beta^{\prime}=\|\mathbf{w}\|^{2}.

The LLRT yields the sufficient statistics

lk=1Ts[T]|[𝐮^s]k|201γ.l_{k}={1\over T}\sum_{s\in[T]}|[\hat{\mathbf{u}}_{s}]_{k}|^{2}\mathop{\gtrless}_{\mathcal{H}^{\prime}0}^{\mathcal{H}^{\prime}1}\gamma. (48)

We study its asymptotic performance. Assume that TT is moderately large, after applying central limit theory, the test statistic distributes as Normal distributions in both hypothesis, i.e.,

lk|H0𝒩(σuk|H02,σuk|H04T),lk|H1𝒩(σuk|H12,σuk|H14T).\begin{split}&l_{k|H0}\sim\mathcal{N}(\sigma_{uk|H0}^{2},{\sigma_{uk|H0}^{4}\over T}),\\ &l_{k|H1}\sim\mathcal{N}(\sigma_{uk|H1}^{2},{\sigma_{uk|H1}^{4}\over T}).\end{split} (49)

Finally, we can relate PdP_{d} and PfaP_{fa} as

Pd=1Φ(βΦ1(1Pfa)+T(βSNRminα)SNRminα),P_{d}=1-\Phi\left({\beta^{\prime}\Phi^{-1}(1-P_{fa})+\sqrt{T}(\beta^{\prime}-SNR_{min}\alpha^{\prime})\over SNR_{min}\alpha^{\prime}}\right), (50)

where Φ()\Phi(\cdot) is the CDF of standard normal distribution. An exemplar ROC curve calculated with (50) is demonstrated in Fig. 12.

Appendix B Proof of Properties of Mapping and Reverse Mapping

B-A Proof of Property 3

Proof.

According to Definition 3, the mapping can be split into two stages: 1) apply modular multiplication to ii, i.e., k=[σi]N[N]k=[\sigma i]_{N}\in[N]; and, 2) convert kk into j[B]j\in[B] with j=kB/Nj=\lfloor kB/N\rfloor.

Similarly, according to Definition 4, the reverse-mapping also can be split into two stages: 1) dilate j[B]j\in[B] into 𝕊{v[N]jNBv<(j+1)NB]}[N]\mathbb{S}\equiv\{v\in[N]\mid j\frac{N}{B}\leq v<(j+1)\frac{N}{B}]\}\subset[N]; and, 2) apply inverse modular multiplication on 𝕊\mathbb{S}, i.e., (j,σ1){[uσ1]Nu𝕊}\mathcal{R}(j,\sigma^{-1})\equiv\{[u\sigma^{-1}]_{N}\mid u\in\mathbb{S}\}.

The first stage of reverse-mapping is the inverse operation of the second stage of mapping, and as a result, k𝕊k\in\mathbb{S}. Hence i=[kσ1]N(j,σ1)i=[k\sigma^{-1}]_{N}\in\mathcal{R}(j,\sigma^{-1}) as desired. ∎

B-B Proof of Property 4

Proof.

We use the two stages of the reverse-mapping in the proof of Property 3. The first stage of the reverse-mapping for ii and jj yields 𝕊1{v[N]iNBv<(i+1)NB]}\mathbb{S}_{1}\equiv\{v\in[N]\mid i\frac{N}{B}\leq v<(i+1)\frac{N}{B}]\} and 𝕊2{v[N]jNBv<(j+1)NB]}\mathbb{S}_{2}\equiv\{v\in[N]\mid j\frac{N}{B}\leq v<(j+1)\frac{N}{B}]\}, respectively. It is not difficult to verify that 𝕊1𝕊2=\mathbb{S}_{1}\cap\mathbb{S}_{2}=\emptyset, provided that iji\neq j.

In what follows, we will prove that the second stage of the reverse-mapping also gives distinct results. Assume that there exists m𝕊1,n𝕊2m\in\mathbb{S}_{1},n\in\mathbb{S}_{2}, such that [mσ1]N=[nσ1]N[m\sigma^{-1}]_{N}=[n\sigma^{-1}]_{N}. Modularly multiply both sides with σ\sigma yields that m=nm=n, which is contradictory with 𝕊1𝕊2=\mathbb{S}_{1}\cap\mathbb{S}_{2}=\emptyset. Hence both stages of the reverse-mapping guarantee the results are distinct for iji\neq j. ∎

Acknowledgment

The authors would like to thank Dr. Predrag Spasojevic and Dr. Anand Sarwate from Rutgers university for initial support of this work. The work of SW was jointly supported by China Scholarship Council and Shanghai Institute of Spaceflight Electronics Technology. The work of VMP was partially supported by an ARO grant W911NF-16-1-0126.

References

  • [1] H. Hassanieh, P. Indyk, D. Katabi, and E. Price, “Nearly optimal sparse fourier transform,” in Proceedings of the forty-fourth annual ACM symposium on Theory of computing, pp. 563–578, ACM, 2012.
  • [2] H. Hassanieh, P. Indyk, D. Katabi, and E. Price, “Simple and practical algorithm for sparse fourier transform,” in Proceedings of the Twenty-third Annual ACM-SIAM Symposium on Discrete Algorithms, SODA ’12, pp. 1183–1194, SIAM, 2012.
  • [3] F. Ong, S. Pawar, and K. Ramchandran, “Fast and efficient sparse 2d discrete fourier transform using sparse-graph codes,” arXiv preprint arXiv:1509.05849, 2015.
  • [4] A. C. Gilbert, P. Indyk, M. Iwen, and L. Schmidt, “Recent developments in the sparse fourier transform: A compressed fourier transform for big data,” Signal Processing Magazine, IEEE, vol. 31, no. 5, pp. 91–100, 2014.
  • [5] B. Ghazi, H. Hassanieh, P. Indyk, D. Katabi, E. Price, and L. Shi, “Sample-optimal average-case sparse fourier transform in two dimensions,” in Communication, Control, and Computing (Allerton), 2013 51st Annual Allerton Conference on, pp. 1258–1265, IEEE, 2013.
  • [6] P. Indyk and M. Kapralov, “Sample-optimal fourier sampling in any constant dimension,” in Foundations of Computer Science (FOCS), 2014 IEEE 55th Annual Symposium on, pp. 514–523, IEEE, 2014.
  • [7] H. Hassanieh, F. Adib, D. Katabi, and P. Indyk, “Faster gps via the sparse fourier transform,” in Proceedings of the 18th annual international conference on Mobile computing and networking, pp. 353–364, ACM, 2012.
  • [8] H. Hassanieh, L. Shi, O. Abari, E. Hamed, and D. Katabi, “Ghz-wide sensing and decoding using the sparse fourier transform,” in INFOCOM, 2014 Proceedings IEEE, pp. 2256–2264, IEEE, 2014.
  • [9] L. Shi, H. Hassanieh, A. Davis, D. Katabi, and F. Durand, “Light field reconstruction using sparsity in the continuous fourier domain,” ACM Transactions on Graphics (TOG), vol. 34, no. 1, p. 12, 2014.
  • [10] A. Rauh and G. R. Arce, “Sparse 2d fast fourier transform,” Proceedings of the 10th International Conference on Sampling Theory and Applications, 2013.
  • [11] P. Boufounos, V. Cevher, A. C. Gilbert, Y. Li, and M. J. Strauss, “What’s the frequency, kenneth?: Sublinear fourier sampling off the grid,” in Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, pp. 61–72, Springer, 2012.
  • [12] E. J. Candè and M. B. Wakin, “An introduction to compressive sampling,” Signal Processing Magazine, IEEE, vol. 25, no. 2, pp. 21–30, 2008.
  • [13] C.-Y. Chen and P. Vaidyanathan, “Compressed sensing in mimo radar,” in Signals, Systems and Computers, 2008 42nd Asilomar Conference on, pp. 41–44, IEEE, 2008.
  • [14] Y. Yu, A. P. Petropulu, and H. V. Poor, “Mimo radar using compressive sampling,” Selected Topics in Signal Processing, IEEE Journal of, vol. 4, no. 1, pp. 146–163, 2010.
  • [15] S. Wang, V. M. Patel, and A. Petropulu, “RSFT: a realistic high dimensional sparse fourier transform and its application in radar signal processing,” in Milcom 2016 Track 1 - Waveforms and Signal Processing., (Baltimore, USA), Avaliable at: http://www.rci.rutgers.edu/ vmp93/Conference_pub/
    MILCOM2016_SFT.pdf, Nov. 2016.
  • [16] F. J. Harris, “On the use of windows for harmonic analysis with the discrete fourier transform,” Proceedings of the IEEE, vol. 66, no. 1, pp. 51–83, 1978.
  • [17] H. L. Van Trees, Optimum array processing: part IV of detection, estimation, and modulation, ch. 5, pp. 349–350. Wiley, New York, 2002.
  • [18] R. B. Ash and C. Doleans-Dade, Probability and measure theory, p. 309. Academic Press, 2000.
  • [19] M. Frigo and S. G. Johnson, “Fftw: An adaptive software architecture for the fft,” in Acoustics, Speech and Signal Processing, 1998. Proceedings of the 1998 IEEE International Conference on, vol. 3, pp. 1381–1384, IEEE, 1998.
  • [20] M. Skolnik, “Systems aspects of digital beam forming ubiquitous radar,” tech. rep., DTIC Document, 2002.
  • [21] D. A. Schade, T. C. Winant, J. Alforque, J. Faul, K. B. Groves, V. Horvatich, M. A. Middione, C. Tarantino, and J. R. Turner, “Fast acting active protection system,” Apr. 10 2007. US Patent 7,202,809.
  • [22] M. I. Skolnik, “Radar handbook,” 1970.
  • [23] P. Stoica and R. L. Moses, Spectral analysis of signals, pp. 49–50. Pearson/Prentice Hall Upper Saddle River, NJ, 2005.
  • [24] H. So, Y. Chan, Q. Ma, and P. Ching, “Comparison of various periodograms for sinusoid detection and frequency estimation,” Aerospace and Electronic Systems, IEEE Transactions on, vol. 35, no. 3, pp. 945–952, 1999.