Second-Order Nonlinearity Estimated and Compensated Diffusion LMS Algorithm: Theoretical Upper Bound, Cramer-Rao Lower bound, and Convergence Analysis
Abstract
In this paper, an algorithm for estimation and compensation of second-order nonlinearity in wireless sensor setwork (WSN) in distributed estimation framework is proposed. First, the effect of second-order nonlinearity on the performance of Diffusion Least Mean Square (DLMS) algorithm is investigated and an upper bound for -norm of the error due to nonlinearity is derived mathematically. Second, mean convergence analysis of the DLMS algorithm in presence of second-order nonlinearity is derived. Third, a distributed algorithm is suggested which consists of extra nonlinearity estimation and compensation units. Moreover, considering the second-order nonlinearity, the Cramer-Rao bound (CRB) for estimating both the unknown vector and nonlinearity coefficient vector is calculated, in which the Fisher information matrix is obtained in a closed-form formula. Simulation results demonstrate the effectiveness of the proposed algorithm in improving the performance of distributed estimation in the presence of nonlinear sensors in a WSN.
keywords:
Distributed estimation, nonlinearity, compensation, second order, diffusion.1 Introduction
The problem of distributed estimation of an unknown vector from linear measurements is a well-known subject in signal processing community which has numerous applications in wireless sensor network (WSN), channel estimation, spectrum estimation, massive MIMO communication, and target tracking problems [1], [2]. The distributed estimation algorithms benefit from the inter-collaboration of sensor nodes. The cooperation strategies for distributed estimation are incremental, consensus, and diffusion approaches [1]. Among them, the diffusion strategy is more versatile due to its simplicity, scalability, and low storage demands.
Many distributed diffusion algorithms are proposed in the literature, e.g., diffusion LMS [3], [4], diffusion LMP [5], [6], diffusion Affine Projection Algorithm (APA) [7], [8], diffusion CMPN [9], and diffusion correntropy [10]-[13], to name a few. Among diffusion algorithms, the Diffusion Least Mean Square (DLMS) algorithm is the basic algorithm which uses mean square error (MSE) as its cost function. There are also numerous variants of the DLMS algorithm, which aim to either reduce the communication load [14], [15], make the algorithm robust against impulsive noise [16], [17], make the algorithm secure with respect to adversaries [18], [19], [20], or in sparse setting [21]. Unfortunately, the performance of the aforementioned algorithms in a WSN deteriorates when the sensors have some nonlinearity effect due e.g. to their power amplifiers. This is because they are designed for the linear measurement model. The main objective of this paper is to make the DLMS algorithm robust against nonlinearities.
In the literature of distributed estimation, there are some works that consider a nonlinear model for the measurements [22]-[28]. In the pioneering work of [22], two distributed algorithms are suggested for estimation in a nonlinear observation model. Moreover, a diffusion based kernel least mean squares (KLMS) is presented in a nonlinear measurement setup [23]. In addition, a distributed estimation algorithm with nonlinear sensors with one bit measurements are proposed in [24]. Besides, [25] suggests two algorithms for estimating the parameters of nonlinear Hammerstein systems with missing data. A distributed nonlinear parameter estimation algorithm is further developed in unbalanced multi-agent networks [26]. Nonlinear model is partially used in [27], in which a method for distributed solution of robust estimation problems is proposed with equality constraints based on the augmented Lagrangian method. [28] discusses both linear and nonlinear models for secure distributed estimation in the presence of attackers in the network.
In this paper, we deal with second-order non-linear model for sensors. It allows to model a linear system which shows some small degree of nonlinearities. The challenges of the second-order nonlinear model is in adaptiveness in which the nonlinear coefficient may change over time. In the proposed solution, the adaptiveness are taken into account. Thanks to the second-order nonlinear model, we can investigate the nonlinearity effect on the performance of a DLMS algorithm. Hence, an upper bound for the error is calculated in the paper. Also, an improved version of DLMS algorithm is suggested which incorporates nonlinearity estimation and compensation units. Further, the Cramer-Rao bound (CRB) is calculated for the distributed estimation problem in the presence of second-order nonlinearity. Simulation results show the benefit of the proposed method especially when there are small nonlinear coefficients.
2 System model and problem formulation
Consider a WSN with sensors (nodes) collecting a scalar measurement , where is the node index and is the time instant. Each sensor contains its own input regression vector . The model of measurements is linear, i.e., , with the unknown vector , where denotes the measurement noise. It is assumed that the sensor equipment has a power amplifier with a second-order nonlinear model. So, if the nonlinear function is then nonlinear measurements are
(1) |
where is the second-order nonlinearity coefficient of th sensor and is the measurement noise which is assumed to be zero-mean Gaussian with variance . The constant term in the above model (1) is omitted since the system is an approximately linear system with a second-order nonlinear term.
The main objective of the distributed estimation problem in the WSN is to estimate the unknown vector using nonlinear measurements and regression vectors of sensors. The other objective is to estimate the nonlinearity of the sensors in the network.
3 The DLMS Algorithm in the presence of nonlinearities
Diffusion algorithms are usually suitable solutions for distributed estimation problems, of which DLMS is the most basic. The two steps of the DLMS algorithm is the adaptation and combination steps. It can be implemented in two ways: Adapt Then Combine (ATC) and Combine Then Adapt (CTA). The ATC version of DLMS is as follows [1]:
(2) |
where , , and denote the neighborhood set of the ’th sensor, the intermediate estimation of ’th sensor in the presence of nonlinearity at time index , and the received intermediate estimation of sensor , respectively. Further, and are the combination coefficients from node to node in the adaptation and combination steps, respectively. The local cost function of node in the DLMS algorithm in presence of nonlinearity is defined as
(3) |
where expectation operator can be neglected and the point estimation can be replaced for expectation.
4 The DLMS Algorithm in the presence of nonlinearities
Diffusion algorithms are usually suitable solutions for distributed estimation problems, of which DLMS is the most basic. The two steps of the DLMS algorithm is the adaptation and combination steps. It can be implemented in two ways: Adapt Then Combine (ATC) and Combine Then Adapt (CTA). The ATC version of DLMS is as follows [1]:
(4) |
where , , and denote the neighborhood set of the ’th sensor, the intermediate estimation of ’th sensor in the presence of nonlinearity at time index , and the received intermediate estimation of sensor , respectively. Further, and are the combination coefficients from node to node in the adaptation and combination steps, respectively. The local cost function of node in the DLMS algorithm in presence of nonlinearity is defined as
(5) |
where expectation operator can be neglected and the point estimation can be replaced for expectation.
5 The Upper bound for the error term due to nonlinearity
In this section, we investigate the effect of second-order nonlinearity on the performance of the DLMS algorithm. We derive an upper bound for the -norm of the error term, which is the difference between the estimated vector after combination step in the presence of nonlinearity and without the nonlinearity. To that end, we write the formula of the intermediate estimation in (4) in the following form
(6) |
where the noise term is neglected in comparison to nonlinear term and is the error vector of the intermediate estimation. In fact, we assume that the noise level is much lower than the nonlinearity term. Then, the intermediate estimations of is received by the node as . So, we have
(7) |
where is the element-wise square of a vector in which is the hadamard operator. Then, the output of combination unit will be
(8) |
where is the new true estimation without nonlinearity. To upper bound the -norm of the error vector of , we write
(9) |
The upper bound for the error term is expressed in the following theorem.
Theorem 1.
The error term is upper bounded by
(10) |
where
(11) |
where is the upper bound of .
Proof.
Neglecting and in (9) for simplicity, and assuming the nonlinear coefficients are small, we can neglect the second order error terms, i.e., . Then, we have the following approximation
(12) |
Using the triangular inequality, we have
(13) |
and we achieve
(14) |
Now, we assume , where is the upper bound for square error of elements. Also, we assume , where is the upper bound of the -norm. We also know that and have linear relationship, i.e., . Thus, we only derive the upper bound . By neglecting the noise term and assuming , we can write
(15) |
Assuming , without loss of generality, we have . Also, using Cauchy-Schuartz inequality, we have . Then, we have
(16) |
Further, we have . After finding the upper bounds and , we derive the upper bound for by applying Cauchy-Schartz inequality. Hence, we have
(17) |
Then, considering the upper bounds, we have
(18) |
where and are the upper bounds of and , respectively. We have , where without loss of generality, we assume that the intermediate estimations are normalized to unity. Similarly, we have . Besides, we have . Hence, following (18) and simplifying the terms, we have
(19) |
Then, substituting in (19), with some simplifications, the proof is achieved. ∎
6 Mean convergence analysis in presence of nonlinearity
In this section, the mean convergence analysis of DLMS in presence of second-order nonlinearity is performed. In the first case, we assume the nonlinearity both in measurements and links. In the second case, we consider the nonlinearity just in measurements.
6.1 Genral-case:nonlinearity both in measurements and links
From the combination step of the DLMS algorithm in presence of second-order nonlinearity, we have
(20) |
where and are given by
(21) |
and
(22) |
where in which we have . So, if we define and , we have
(23) |
Then, if we define , , and , then (23) can be written in the following form
(24) |
In the appendix 1, the , , and are calculated as
(25) |
,
(26) |
and
(27) |
where and . In appendix 2, the and are computed in Appendix 2.
6.2 Special-case:nonlinearity only in measurements
In this part, the mean convergence analysis is performed when the only nonlinearity is in measurements. In this case, we can write
(30) |
Then, (30) can be written as
(31) |
By taking expectation from both sides of (31), we have
(32) |
Now, (32) can be written as a recursion without bias term, as follows:
(33) |
where .
7 The Proposed Algorithm
DLMS algorithm performance is degraded by nonlinearity. To improve the performance of the DLMS algorithm in the presence of nonlinearity, we propose to estimate the nonlinear coefficients and then compensate their effects. This process is adaptive and online as it shows its benefits in comparison to a pre-calibration process. We call the proposed algorithm second-order nonlinearity estimated and compensated DLMS (SONEC-DLMS) algorithm. It consists of five different steps. The details of steps are:
-
1.
Nonlinear coefficient estimation: If we define the coefficient vector of node as , we estimate this vector at node by suggesting the following cost function as:
(34) where is the linear ground truth measurement vector of node with incorporating the measurement noise and is estimated in the second step of the algorithm, and is the nonlinear coefficients estimated by node . To find , we use a steepest descent of by assuming is known. So, we have
(35) where , , and .
-
2.
True measurement estimation: In this step, since we estimate from , i.e., the second-order nonlinear equation, we can equivalently call this step compensation of nonlinear measurements. We use the second-order equation . So, neglecting the other incorrect solution, the correct solution is
(36) -
3.
Adaptation step: This step is equivalent to a classical adaptation step in DLMS algorithm, i.e.,
(37) where .
-
4.
Compensation of nonlinearity in the intermediate estimation: After exchanging the intermediate estimations of , the received intermediate estimations, i.e., , should be compensated by the relation of second order nonlinear equation. Similar to step 2, we write
(38) -
5.
Combination step: In this step, we update the final estimation of node as .
Hence, the SONEC-DLMS algorithm is a five-step diffusion algorithm having three extra steps compared to DLMS algorithm. These three steps are, one step for estimation of nonlinear coefficients and, two other steps for compensation of nonlinearity of measurements and intermediate estimations. We call this algorithm fully-distributed SONEC-DLMS algorithm. The computational complexity of the proposed SONEC-DLMS in terms of number of addition, multiplication, and nonlinear operator of square root, in comparison to DLMS algorithm is depicted in Table 1. It is seen that the computational complexity of the proposed algorithm is approximately three times that of the DLMS algorithm.
Algorithm | Add | Multiplication | Nonlinear |
---|---|---|---|
DLMS | |||
SONEC-DLMS |
Nonlinear: Nonlinear operators such as square root. |
To further improve the performance of the SONEC-DLMS algorithm, we propose a semi-distributed SONEC-DLMS algorithm, in which the nonlinear coefficients are estimated in a centralized manner by a fusion center by using training data. This version of the proposed algorithm, which can be considered as a combination of distributed and centralized algorithms, contains only four steps and the nonlinearity estimation step is done separately in a centralized manner. As we will see in the simulation results, the performance of this semi-distributed algorithm is close to that of DLMS without nonlinearity.
8 Cramer-Rao bound
In this section, the Cramer-Rao bound for estimating the nonlinear coefficient vector and unknown vector is derived. Let us define an parameter vector . We intend to find the CRB of estimated based on the measurements . All the measurements are defined in a matrix where is the total observations of node . From (1), we have , where represents the Gaussian distribution with mean and variance . The Fisher Information Matrix (FIM) of measurements for estimating is calculated as , where , , and .
To further proceed and to calculate the likelihood , we write the observation vector as
(39) |
where and . Assuming the independence of measurement noises in different nodes , we have . Hence, the log-likelihood can be written as
(40) |
So, the partial derivative is . To further proceed, we define .
Let , then some calculations lead to
(41) |
Using (41) and taking expectations and considering that , we have
(42) |
where . Some calculations show that we have
(43) |
where . Similar calculations can lead to
(44) |
(45) |
where
(46) |
9 Simulation Results
In this section, some experiments are performed to investigate the performance of the proposed SONEC-DLMS algorithm in a distributed network with nonlinear sensors. The simulation setup is as follows. The WSN consists of sensors. The network is selected the same as the one introduced in [19]. The sensors of WSN are collected a second order nonlinear measurement of a unknown vector with L=20. The unknown vector elements and the elements of the regression vectors are chosen from normal distribution. For the nonlinear coefficients, we use a uniform random variable . Unless otherwise stated, we use in the simulations. The proposed distributed estimation algorithm aims to estimate the unknown vector in an adaptive manner. For the background noise , we use a zero-mean Gaussian distribution with standard deviation equal to 0.045. The performance metric for evaluating the performance of the proposed algorithm is the mean square deviation (MSD) criterion defined as . The number of Monte Carlo simulations is 100 independent runs and the results are averaged over all runs. The combination coefficients and are chosen by the uniform policy [1].
The performance of the proposed SONEC-DLMS algorithm with its two different versions which are fully-distributed and semi-distributed is investigated. Furthermore, we have included a version of the proposed algorithm that utilizes compensation solely in the combination step. However, we have excluded another variant of the algorithm that applies compensation in the adaptation step but not in the combination step, as it failed to converge in the simulations. The step-sizes and have been carefully selected to minimize the final mean square deviation (MSD), with values of and . The MSD of estimating the unknown vector versus iteration number is depicted in Fig. 1. As it can be seen, the fully distributed SONEC-DLMS algorithm performs better than DLMS by at least 7dB. It shows that the performance of the semi-distributed SONEC-DLMS algorithm is close to that of the DLMS without nonlinearity. It demonstrates that there is a gap of at least 10dB between the CRB and the SONEC-DLMS. It also shows that the upper-bound is not tight, but is less than . Moreover, the MSD of estimating the nonlinear coefficient vector versus iteration number is shown in Fig. 2. It shows a large gap between the performance of SONEC-DLMS algorithms and CRB.


10 Conclusion
In this paper, a solution for improving the performance of distributed estimation in the presence of nonlinearities was provided. The solution is to use a nonlinearity estimation and nonlinearity compensation units. Moreover, an upper bound for the error due to nonlinearity is obtained. Also, the CRB of the problem of distributed estimation in the presence of second-order nonlinearity was calculated. Simulation results show the effectiveness of the proposed algorithm. The future work could be working on the low resolution messages between sensors [29].
Appendix 1
To calculate , we can write
(47) |
To calculate , we nominate and for simplicity. So, we have
(48) |
since and for are uncorrelated. Hence, we have
(49) |
Therefore, we reach to (25). To calculate , we can write
(50) |
where . Then, we have
(51) |
where the second and third term in (51) are zero. So, we have
(52) |
where we define and for simplicity. Then, we have
(53) |
where we have
(54) |
Hence, simple calculations show that
(55) |
Using the whiteness of , some simple calculations show that . Hence, we reach to (26). To calculate , we have
(56) |
The second term in (56) is zero if we assume the uncorelatedness of and and also since we proved that in the current appendix. So, we have
(57) |
Appendix 2
To compute , we have
(58) |
To calculate in (58), some calculations show that
(59) |
Now, is equal to
(60) |
Therefore, from (58), we have
(61) |
To calculate , we can write
(62) |
So, we have
(63) |
To calculate
(64) |
for simplicity, it is re-written as
(65) |
Then, some calculations lead to
(66) |
It is easy to show that if , we have . Also, for , we have
(67) |
Finally, from (63), we have
(68) |
References
- [1] A. H. Sayed, Adaptation, Learning and Optimization over networks, Foundations and Trends in Machine Learning, 2014.
- [2] A. Shirazinia, et al, “Massive MIMO for Decentralized Estimation of a Correlated Source,” IEEE Trans. on Signal Proc., vol. 64, no. 10, pp. 2499–2512, May 2016.
- [3] C. G. Lopes, and A. H. Sayed, “Diffusion least-mean squares over adaptive networks: Formulation and performance analysis,” IEEE Trans. on Signal Proc., vol. 56, pp. 3122–3136, 2008.
- [4] F. S. Cattivelli, and A. H. Sayed, “Diffusion LMS strategies for distributed estimation,” IEEE Trans. on Signal Proc., vol. 58, pp. 1035–1048, 2010.
- [5] F. Wen, “Diffusion Least Mean P-power algorithms for distributed estimation in alpha-stable noise environments ,” Electron. Lett., vol. 49, no. 21, pp. 1355–1356, 2013.
- [6] L. Lu, H. Zhao, W. Wang, and Y. Yu, “Performance analysis of the robust diffusion normalized least mean p-power algorithm,” IEEE Trans. on Circuit and Systems-II: Express Briefs., vol. 65, no. 12, pp. 2047–2051, Dec 2018.
- [7] M. Shams Esfand Abadi, and M. S. Shafiee, “Distributed Estimation Over an Adaptive Diffusion Network Based on the Family of Affine Projection Algorithms,” IEEE Transactions on Signal and Information Processing over Networks, vol. 5, no. 2, pp. 234–247, June 2019.
- [8] H. Shiri, M. A. Tinati, M. Codreanu, and S. Daneshvar, “Distributed Sparse Diffusion Estimation based on Set Membership and Affine Projection Algorithm,” DSP Signal Processing., vol. 73, pp. 47–61, Feb 2018.
- [9] M. Korki, and H. Zayyani, “Weighted diffusion continuous mixed p-norm algorithm for distributed estimation in non-uniform noise environment,” Elsevier Signal Processing., vol. 164, pp. 225–233, Nov 2019.
- [10] W. Ma, B. Chen, J. Duan, and H. Zhao, “Diffusion maximum correntropy criterion algorithms for robust distributed estimation,” Digital Signal Processing., vol. 58, pp. 10-16, 2016.
- [11] F. Chen, et al, “Diffusion generalized maximum correntropy criterion algorithm for distributed estimation over multitask network,” Digital Signal Processing., vol. 81, pp. 16-25, Oct 2018.
- [12] Y. He, F. Wang, S. Wang, P. Ren, and B. Chen, “Maximum total correntropy diffusion adaptation over networks with noisy links,” IEEE Trans. on Circuit and Systems-II: Express Briefs., vol. 66, no. 2, pp. 307–311, Feb 2019.
- [13] V. C. Gogineni, S. P. Talebi, S. Werner, and D. P. Mandic, “Fractional-Order Correntropy Adaptive Filters for Distributed Processing of a-Stable Signals,” IEEE Signal Processing Letters, vol. 27, pp. 1884–1888, 2020.
- [14] R. Arablouei, S. Werner, Y. Huang, and K. Dogancay, “Distributed Least Mean-Square Estimation With Partial Diffusion,” IEEE Tran. Signal Processing, vol. 62, no. 2, pp. 472–484, 2014.
- [15] H. Zayyani, “Communication reducing diffusion LMS robust to impulsive noise using smart selection of communication nodes,” Circuit, System, and Signal Processing, vol. 41, no. 3, pp. 1788–1802, 2022.
- [16] H. Zayyani, “Robust minimum disturbance diffusion LMS for distributed estimation,” IEEE Transactions on Circuits and Systems II: Express Briefs, vol. 68, no. 1, pp. 521–525, 2020.
- [17] H. Zayyani, and A. Javaheri, “A Robust Generalized Proportionate Diffusion LMS Algorithm for Distirbuted Estimation,” IEEE Trans on Circuits and systems-II: Express Briefs., vol. 68, no. 4, pp. 1552-1556 , April 2021.
- [18] H. Chang, and W. Li, “Correction-based diffusion LMS algorithms for secure distributed estimation under attacks,” Digital Signal Processing, vol. 102, July 2020.
- [19] H. Zayyani, F. Oruji, and I. Fijalkow, “An Adversary-Resilient Doubly Compressed Diffusion LMS Algorithm for Distributed Estimation,” Circuit, System, and Signal Processing, vol. 41, pp. 6182–6205, 2022.
- [20] H. Zayyani, M. Korki, “Adaptive-Width Generalized Correntropy Diffusion Algorithm for Secure Distributed Estimation,” IEEE Transactions on Circuits and Systems II: Express Briefs, 2023.
- [21] S. Modalavalasa, U. K. Sahoo, and A. K. Sahoo, “Robust non-parametric sparse distributed regression over wireless networks,” DSP Signal Processing., vol. 104, Sep 2020.
- [22] S. Kar, J. M. F. Moura, and K. Ramanan, “Distributed Parameter Estimation in Sensor Networks: Nonlinear Observation Models and Imperfect Communication,” IEEE Trans on Information Theory., vol. 58, no. 6, pp. 3575-3604 , June 2012.
- [23] S. Chouvardas, and M. Draief, “A diffusion kernel LMS algorithm for nonlinear adaptive networks,” ICASSP 2016., Shanghai, China, 2016.
- [24] H. Zayyani, R. Sari, and M. Korki, “A Distributed 1-bit Compressed Sensing Algorithm for Nonlinear Sensors With a Cramer Rao Bound,” IEEE Communication Letters., vol. 21, no. 12, pp. 2626-2629 , Dec 2017.
- [25] S. Chen, and Y. Liu, “Robust Distributed Parameter Estimation of Nonlinear Systems With Missing Data Over Networks,” IEEE Transactions on Aerospace and Electronic Systems., vol. 56, no. 3, pp. 2228-2244 , June 2020.
- [26] M. Meng, and X. Li, “Distributed Nonlinear Estimation Over Unbalanced Directed Networks,” IEEE Transactions Signal Processing., vol. 68, pp. 6212-6223 , 2020.
- [27] S. C. Chan, H. C. Wu, C. H. Ho, and L. Zhang, “An Augmented Lagrangian Approach for Distributed Robust Estimation in Large-Scale Systems,” IEEE Systems Journal., vol. 13, no. 3, pp. 2986-2997 , Sep 2019.
- [28] M. Meng, X. Li, and G. Xiao, “Distributed Estimation Under Sensor Attacks: Linear and Nonlinear Measurement Models,” IEEE Transactions on Signal and Information Processing over Networks, vol. 7, pp. 156–165, 2021.
- [29] D. Ciuonzo, et al, “Bandwidth-Constrained Decentralized Detection of an Unknown Vector Signal via Multisensor Fusion,” IEEE Transactions on Signal and Information Processing over Networks, vol. 6, pp. 744–758, 2020.