This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Study of Robust Adaptive Power Allocation Techniques for Rate Splitting based MU-MIMO systems

A. R. Flores and R. C. de Lamare This work was partially supported by the National Council for Scientific and Technological Development (CNPq), FAPERJ, FAPESP, and CGI.A. R. Flores is with the Centre for Telecommunications Studies, Pontifical Catholic University of Rio de Janeiro, Rio de Janeiro 22451-900, Brazil (e-mail: andre.flores@cetuc.puc-rio.br).R. C. de Lamare is with the Centre for Telecommunications Studies, Pontifical Catholic University of Rio de Janeiro, Rio de Janeiro 22451-900, Brazil, and also with the Department of Electronic Engineering, University of York, York YO10 5DD, U.K. (e-mail: delamare@cetuc.puc-rio.br).
Abstract

Rate splitting (RS) systems can better deal with imperfect channel state information at the transmitter (CSIT) than conventional approaches. However, this requires an appropriate power allocation that often has a high computational complexity, which might be inadequate for practical and large systems. To this end, adaptive power allocation techniques can provide good performance with low computational cost. This work presents novel robust and adaptive power allocation technique for RS-based multiuser multiple-input multiple-output (MU-MIMO) systems. In particular, we develop a robust adaptive power allocation based on stochastic gradient learning and the minimization of the mean-square error between the transmitted symbols of the RS system and the received signal. The proposed robust power allocation strategy incorporates knowledge of the variance of the channel errors to deal with imperfect CSIT and adjust power levels in the presence of uncertainty. An analysis of the convexity and stability of the proposed power allocation algorithms is provided, together with a study of their computational complexity and theoretical bounds relating the power allocation strategies. Numerical results show that the sum-rate of an RS system with adaptive power allocation outperforms RS and conventional MU-MIMO systems under imperfect CSIT.

Index Terms:
Rate splitting, multiuser MIMO, power allocation, ergodic sum rate.

I Introduction

Wireless communications systems employing multiple antennas have the advantage of increasing the overall throughput without increasing the required bandwidth. For this reason, multiple-antenna systems are at the core of several wireless communications standards such as WiFi, Long Term Evolution (LTE) and the fifth generation (5G). However, such wireless systems suffer from multiuser interference (MUI). In order to mitigate MUI, transmit processing techniques have been employed in the downlink (DL), allowing accurate recovery of the data at the receivers. In general, a precoding technique maps the symbols containing the information to the transmit antennas so that the information arrives at the receiver with reduced levels of MUI. Due to its benefits, linear [1, 2, 3, 4, 5, 6] and non-linear [7, 8, 9] precoding techniques have been extensively reported in the literature. The design of effective precoders demands very accurate channel state information at the transmitter (CSIT), which is an extremely difficult task to accomplish in actual wireless systems. Hence, the transmitter typically only has access to partial or imperfect CSIT. As a result, the precoder cannot handle MUI as expected, resulting in residual MUI at the receiver. This residual MUI can degrade heavily the performance of wireless systems since it scales with the transmit power employed at the base station (BS) [10].

I-A Prior and related work

In this context, Rate Splitting (RS) has emerged as a novel approach that is capable of dealing with CSIT imperfection [11] in an effective way. RS was initially proposed in [12] to deal with interference channels [13], where independent transmitters send information to independent receivers [14]. Since then, several studies have found that RS outperforms conventional schemes such as conventional precoding in spatial division multiple access (SDMA), power-domain Non-Orthogonal Multiple Access (NOMA) [15] and even dirty paper coding (DPC) [16]. Interestingly, it turns out that RS constitutes a generalized framework which has as special cases other transmission schemes such SDMA, NOMA and multicasting [17, 18, 19, 20]. The main advantage of RS is its capability to partially decode interference and partially treat interference as noise. To this end, RS splits the message of one or several users into a common message and a private message. The common message must be decoded by all the users that employ successive interference cancellation [21, 22, 23, 24, 25]. On the other hand, the private messages are decoded only by their corresponding users.

RS schemes have been shown to enhance the performance of wireless communication systems. In [26] RS was extended to the broadcast channel (BC) of multiple-input single-output (MISO) systems, where it was shown that RS provides gains in terms of Degrees-of-Freedom (DoF) with respect to conventional multiuser multiple-input multiple-output (MU-MIMO) schemes under imperfect CSIT. Later in [27], the DoF of a MIMO BC and IC was characterized. RS has eventually been shown in [28] to achieve the optimal DoF region when considering imperfect CSIT, outperforming the DoF obtained by SDMA systems, which decays in the presence of imperfect CSIT.

Due to its benefits, several wireless communications deployments with RS have been studied. RS has been employed in MISO systems along with linear precoders [29, 30] in order to maximize the sum-rate performance under perfect and imperfect CSIT assumptions. Another approach has been presented in [31] where the max-min fairness problem has been studied. In [32] a K-cell MISO IC has been considered and the scheme known as topological RS presented. The topological RS scheme transmits multiple layers of common messages, so that the common messages are not decoded by all users but by groups of users. RS has been employed along with random vector quantization in [33] to mitigate the effects of the imperfect CSIT caused by finite feedback. In [34, 35], RS with common stream combining techniques has been developed in order to exploit multiple antennas at the receiver and to improve the overall sum-rate performance. A successive null-space precoder, that employs null-space basis vectors to adjust the inter-user-interference at the receivers, is proposed in [36]. The optimization of the precoders along with the transmission of multiple common streams was considered in [37]. In [38], RS with joint decoding has been explored. The authors of [38] devised precoding algorithms for an arbitrary number of users along with a stream selection strategy to reduce the number of precoded signals.

Along with the design of the precoders, power allocation is also a fundamental part of RS-based systems. The benefits of RS are obtained only if an appropriate power allocation for the common stream is performed. However, the power allocation problem in MU-MIMO systems is an NP-hard problem [39, 40], and the optimal solution can be found at the expense of an exponential growth in computational complexity. Therefore, suboptimal approaches that jointly optimize the precoder and the power allocation have been developed. Most works so far employ exhaustive search or complex optimization frameworks. These frameworks rely on the augmented WMMSE [16, 41, 42, 43, 37], which is an extension of the WMMSE proposed in [44]. This approach requires also an alternating optimization, which further increases the computational complexity. A simplified approach can be found in [45], where closed-form expressions for RS-based massive MIMO systems are derived. However, this suboptimal solution is more appropiate for massive MIMO deployments. The high complexity of most optimal approaches makes them impractical to implement in large or real-time systems. For this reason, there is a strong demand for cost-effective power allocation techniques for RS systems.

I-B Contributions

In this paper, we present novel efficient robust and adaptive power allocation techniques [46] for RS-based MU-MIMO systems. In particular, we develop a robust adaptive power allocation (APA-R) strategy based on stochastic gradient learning [47, 48, 49, 50] and the minimization of the mean-square error (MSE) between the transmitted common and private symbols of the RS system and the received signal. We incorporate knowledge of the variance of the channel errors to cope with imperfect CSIT and adjust power levels in the presence of uncertainty. When the knowledge of the variance of the channel errors is not exploited the proposed APA-R becomes the proposed adaptive power allocation algorithm (APA). An analysis of the convexity and stability of the proposed power allocation algorithms along with a study of their computational complexity and theoretical bounds relating the power allocation strategies are developed. Numerical results show that the sum-rate of an RS system employing adaptive power allocation outperforms conventional MU-MIMO systems under imperfect CSIT assumption. The contributions of this work can be summarized as:

  • Cost-effective APA-R and APA techniques for power allocation are proposed based on stochastic gradient recursions and knowledge of the variance of the channel errors for RS-based and standard MU-MIMO systems.

  • An analysis of convexity and stability of the proposed power allocation techniques along with a bound on the MSE of APA and APA-R and a study of their computational complexity.

  • A simulation study of the proposed APA and APA-R, and the existing power allocation techniques for RS-based and standard MU-MIMO systems.

I-C Organization

The rest of this paper is organized as follows. Section II describes the mathematical model of an RS-based MU-MIMO system. In Section III the proposed APA-R technique is presented, the proposed APA approach is obtained as a particular case and sum-rate expressions are derived. In Section IV, we present an analysis of convexity and stability of the proposed APA and APA-R techniques along with a bound on the MSE of APA and APA-R and a study of their computational complexity. Simulation results are illustrated and discussed in Section V. Finally, Section VI presents the conclusions of this work.

I-D Notation

Column vectors are denoted by lowercase boldface letters. The vector 𝐚i,\mathbf{a}_{i,*} stands for the iith row of matrix 𝐀\mathbf{A}. Matrices are denoted by uppercase boldface letters. Scalars are denoted by standard letters. The superscript ()T\left(\cdot\right)^{\text{T}} represents the transpose of a matrix, whereas the notation ()H\left(\cdot\right)^{H} stands for the conjugate transpose of a matrix. The operators \lVert\cdot\rVert, and 𝔼x[]\mathbb{E}_{x}\left[\cdot\right] represent the Euclidean norm, and the expectation w.r.t the random variable xx, respectively. The trace operator is given by tr()\text{tr}\left(\cdot\right). The Hadamard product is denoted by \odot. The operator diag(𝐚)\text{diag}\left(\mathbf{a}\right) produces a diagonal matrix with the coefficients of 𝐚\mathbf{a} located in the main diagonal.

II System Model

Let us consider the RS-based MU-MIMO system architecture depicted in Fig. 1, where the BS is equipped with NtN_{t} antennas, serves KK users and the kkth UE is equipped with NkN_{k} antennas. Let us denote by NrN_{r} the total number of receive antennas. Then, Nr=k=1KNkN_{r}=\sum_{k=1}^{K}N_{k}. For simplicity, the message intended for the kkth user is split into a common message and a private message. Then, the messages are encoded and modulated. The transmitter sends one common stream and a total of MM private streams, with MNrM\leq N_{r}. The set k\mathcal{M}_{k} contains the MkM_{k} private streams, that are intended for the user kk, where MkNkM_{k}\leq N_{k}. It follows that M=k=1KMkM=\sum_{k^{=}1}^{K}M_{k}.

Refer to caption
Figure 1: RS MU-MIMO architecture.

The vector 𝐬(RS)=[sc,𝐬pT]TM+1\mathbf{s}^{\left(\text{RS}\right)}=\left[s_{c},\mathbf{s}_{p}^{\text{T}}\right]^{\text{T}}\in\mathbb{C}^{M+1}, which is assumed i.i.d. with zero mean and covariance matrix equal to the identity matrix, contains the information transmitted to all users, where scs_{c} is the common symbol and 𝐬p=[𝐬1T,𝐬2T,,𝐬KT]T\mathbf{s}_{p}=\left[\mathbf{s}_{1}^{\text{T}},\mathbf{s}_{2}^{\text{T}},\cdots,\mathbf{s}_{K}^{\text{T}}\right]^{\text{T}} contains the private symbols of all users. Specifically, the vector 𝐬kMk\mathbf{s}_{k}\in\mathbb{C}^{M_{k}} contains the private streams intended for the kkth user. The system is subject to a transmit power constraint given by 𝔼[𝐱(RS)2]Etr\mathbb{E}\left[\lVert\mathbf{x}^{\left(\text{RS}\right)}\rVert^{2}\right]\leq E_{tr}, where 𝐱(RS)Nt\mathbf{x}^{\left(\text{RS}\right)}\in\mathbb{C}^{N_{t}} is the transmitted vector and EtrE_{tr} denotes the total available power. The transmitted vector can be expressed as follows:

𝐱(RS)=\displaystyle\mathbf{x}^{\left(\text{RS}\right)}= 𝐏(RS)𝐀(RS)𝐬(RS)=acsc𝐩c+m=1Mamsm𝐩m,\displaystyle\mathbf{P}^{\left(\text{RS}\right)}\mathbf{A}^{\left(\text{RS}\right)}\mathbf{s}^{\left(\text{RS}\right)}=a_{c}s_{c}\mathbf{p}_{c}+\sum_{m=1}^{M}a_{m}s_{m}\mathbf{p}_{m}, (1)

where 𝐀(RS)(M+1)×(M+1)\mathbf{A}^{\left(\text{RS}\right)}\in\mathbb{R}^{\left(M+1\right)\times\left(M+1\right)} represents the power allocation matrix and 𝐏(RS)=[𝐩c,𝐩1,,𝐩K]Nt×(M+1)\mathbf{P}^{\left(\text{RS}\right)}=[\mathbf{p}_{c},\mathbf{p}_{1},\cdots,\mathbf{p}_{K}]\in\mathbb{C}^{N_{t}\times\left(M+1\right)} is used to precode the vector of symbols 𝐬(RS)\mathbf{s}^{\left(\text{RS}\right)}. Specifically, 𝐀RS=diag(𝐚(RS))\mathbf{A}^{\text{RS}}=\text{diag}\left(\mathbf{a}^{\left(\text{RS}\right)}\right) and 𝐚(RS)=[ac,a1,,aM]TM+1\mathbf{a}^{\left(\text{RS}\right)}=\left[a_{c},a_{1},\cdots,a_{M}\right]^{\text{T}}\in\mathbb{R}^{M+1}, where aca_{c} denotes the power allocated to the common stream and aka_{k} allocates power to the kkth private stream. Without loss of generality, we assume that the columns of the precoders are normalized to have unit norm.

The model established leads us to the received signal described by

𝐲=𝐇𝐏(RS)diag(𝐚(RS))𝐬(RS)+𝐧,\mathbf{y}=\mathbf{H}\mathbf{P}^{\left(\text{RS}\right)}\text{diag}\left(\mathbf{a}^{\left(\text{RS}\right)}\right)\mathbf{s}^{\left(\text{RS}\right)}+\mathbf{n}, (2)

where 𝐧=[𝐧1T,𝐧2T,,𝐧KT]TNr\mathbf{n}=\left[\mathbf{n}_{1}^{\text{T}},\mathbf{n}_{2}^{\text{T}},\cdots,\mathbf{n}_{K}^{\text{T}}\right]^{\text{T}}\in\mathbb{C}^{N_{r}} represents the uncorrelated noise vector, which follows a complex normal distribution, i.e., 𝐧𝒞𝒩(𝟎,σn2𝐈)\mathbf{n}\sim\mathcal{CN}\left(\mathbf{0},\sigma_{n}^{2}\mathbf{I}\right). We assume that the noise and the symbols are uncorrelated, which is usually the case in real systems. The matrix 𝐇=[𝐇1T,𝐇2T,,𝐇KT]TNr×Nt\mathbf{H}=\left[\mathbf{H}_{1}^{\text{T}},\mathbf{H}_{2}^{\text{T}},\cdots,\mathbf{H}_{K}^{\text{T}}\right]^{\text{T}}\in\mathbb{C}^{N_{r}\times N_{t}} denotes the channel between the BS and the user terminals. Specifically, 𝐧k\mathbf{n}_{k} denotes the noise affecting the kkth user and the matrix 𝐇kNk×Nt\mathbf{H}_{k}\in\mathbb{C}^{N_{k}\times N_{t}} represents the channel between the BS and the kkth user. The imperfections in the channel estimate are modelled by the random matrix 𝐇~\tilde{\mathbf{H}}. Each coefficient of 𝐇~\tilde{\mathbf{H}} follows a Gaussian distribution with variance equal to σe,i2\sigma_{e,i}^{2} and 𝔼[𝐡~i,H𝐡~i,]=σe2𝐈i[1,Nr]\mathbb{E}\left[\tilde{\mathbf{h}}_{i,*}^{H}\tilde{\mathbf{h}}_{i,*}\right]=\sigma_{e}^{2}\mathbf{I}\quad\forall i\in\left[1,N_{r}\right]. Then, the channel matrix can be expressed as 𝐇=𝐇^+𝐇~\mathbf{H}=\hat{\mathbf{H}}+\tilde{\mathbf{H}}, where the channel estimate is given by 𝐇^\hat{\mathbf{H}}. From (2) we can obtain the received signal of user kk, which is given by

𝐲k=\displaystyle\mathbf{y}_{k}= acsc𝐇k𝐩c+𝐇kikaisi𝐩i+𝐇kl=1lkKjlajsj𝐩j+𝐧k.\displaystyle a_{c}s_{c}\mathbf{H}_{k}\mathbf{p}_{c}+\mathbf{H}_{k}\sum_{i\in\mathcal{M}_{k}}a_{i}s_{i}\mathbf{p}_{i}+\mathbf{H}_{k}\sum\limits_{\begin{subarray}{c}l=1\\ l\neq k\end{subarray}}^{K}\sum\limits_{j\in\mathcal{M}_{l}}a_{j}s_{j}\mathbf{p}_{j}+\mathbf{n}_{k}. (3)

Note that the RS architecture contains the conventional MU-MIMO as a special case where no message is split and therefore aca_{c} is set to zero. Then, the model boils down to the model of a conventional MU-MIMO system, where the received signal at the kkth user is given by

𝐲k=𝐇kikaisi𝐩i+𝐇kl=1lkKjlajsj𝐩j+𝐧k\mathbf{y}_{k}=\mathbf{H}_{k}\sum_{i\in\mathcal{M}_{k}}a_{i}s_{i}\mathbf{p}_{i}+\mathbf{H}_{k}\sum\limits_{\begin{subarray}{c}l=1\\ l\neq k\end{subarray}}^{K}\sum\limits_{j\in\mathcal{M}_{l}}a_{j}s_{j}\mathbf{p}_{j}+\mathbf{n}_{k} (4)

In what follows, we will focus on the development of power allocation techniques that can cost-effectively compute aca_{c} and aj,j=1,,Ka_{j},j=1,\ldots,K.

III Proposed Power Allocation Techniques

In this section, we detail the proposed power allocation techniques. In particular, we start with the derivation of the ARA-R approach and then present the APA technique as a particular case of the APA-R approach.

III-A Robust Adaptive Power Allocation

Here, a robust adaptive power allocation algorithm, denoted as APA-R, is developed to perform power allocation in the presence of imperfect CSIT. The key idea is to incorporate knowledge about the variance of the channel uncertainty [51, 52] into an adaptive recursion to allocate the power among the streams. The minimization of the MSE between the received signal and the transmitted symbols is adopted as the criterion to derive the algorithm.

Let us consider the model established in (2) and denote the fraction of power allocated to the common stream by the parameter δ\delta, i.e., ac2=δEtra_{c}^{2}=\delta E_{tr}. It follows that the available power for the private streams is reduced to (1δ)Etr\left(1-\delta\right)E_{tr}. We remark that the length of 𝐬(RS)\mathbf{s}^{\left(\text{RS}\right)} is greater than that of 𝐲(RS)\mathbf{y}^{\left(\text{RS}\right)} since the common symbol is superimposed to the private symbols. Therefore, we consider the vector 𝐲=𝐓𝐲(RS)\mathbf{y}^{\prime}=\mathbf{T}\mathbf{y}^{\left(\text{RS}\right)}, where 𝐓(M+1)×M\mathbf{T}\in\mathbb{R}^{\left(M+1\right)\times M} is a transformation matrix employed to ensure that the dimensions of 𝐬(RS)\mathbf{s}^{\left(\text{RS}\right)} and 𝐲(RS)\mathbf{y}^{\left(\text{RS}\right)} match, and is given by

𝐓=[111100010001].\mathbf{T}=\begin{bmatrix}1&1&\cdots&1\\ 1&0&\cdots&0\\ 0&1&\cdots&0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0&\cdots&1\end{bmatrix}. (5)

All the elements in the first row of matrix 𝐓\mathbf{T} are equal to one in order to take into account the common symbol obtained at all receivers. As a result we obtain the combined receive signal of all users. It follows that

𝐲=[ycy1yM]=[i=1Myiy1yM],\mathbf{y}^{\prime}=\begin{bmatrix}y_{c}\\ y_{1}\\ \vdots\\ y_{M}\end{bmatrix}=\begin{bmatrix}\sum_{i=1}^{M}y_{i}\\ y_{1}\\ \vdots\\ y_{M}\end{bmatrix}, (6)

where the received signal at the iith antenna is described by

yi=acsc(𝐡^i,+𝐡~i,)𝐩c+j=1Majsj(𝐡^i,+𝐡~i,)𝐩j+ni.y_{i}=a_{c}s_{c}\left(\hat{\mathbf{h}}_{i,*}+\tilde{\mathbf{h}}_{i,*}\right)\mathbf{p}_{c}+\sum_{j=1}^{M}a_{j}s_{j}\left(\hat{\mathbf{h}}_{i,*}+\tilde{\mathbf{h}}_{i,*}\right)\mathbf{p}_{j}+n_{i}. (7)

Let us now consider the proposed robust power allocation problem for imperfect CSIT scenarios. By including the error of the channel estimate, the robust power allocation problem can be formulated as the constrained optimization given by

min𝐚𝔼[𝐬(RS)𝐲(𝐇)2|𝐇^]s.t.tr(𝐏(RS)diag(𝐚(RS)𝐚(RS))𝐏(RS)H)=Etr,\begin{gathered}\min_{\mathbf{a}}\mathbb{E}\left[\lVert\mathbf{s}^{\left(\text{RS}\right)}-\mathbf{y}^{\prime}\left(\mathbf{H}\right)\rVert^{2}|\hat{\mathbf{H}}\right]\\ \text{s.t.}~{}~{}\text{tr}\left(\mathbf{P}^{\left(\text{RS}\right)}\text{diag}\left(\mathbf{a}^{\left(\text{RS}\right)}\odot\mathbf{a}^{\left(\text{RS}\right)}\right)\mathbf{P}^{\left(\text{RS}\right)^{H}}\right)=E_{tr},\end{gathered} (8)

which can be solved by first relaxing the constraint, using an adaptive learning recursion and then enforcing the constraint.

In this work, we choose the MSE as the objective function due to its convex property and mathematical tractability, which help to find an appropriate solution through algorithms. The objective function is convex as illustrated by Fig. 2 and analytically shown in Section IV. In Fig. 2, we plot the objective function using two precoders, namely the zero-forcing (ZF) and the matched filter (MF) precoders [2], where three private streams and one common stream are transmitted and the parameter δ\delta varies with uniform power allocation across private streams.

Refer to caption
Figure 2: Objective function considering a MU-MIMO system with Nt=3Nt=3, K=3K=3, and σn2=1\sigma_{n}^{2}=1

To solve (8) we need to expand the terms and evaluate the expected value. Let us consider that the square error is equal to ε=𝐬(RS)𝐲2\varepsilon=\lVert\mathbf{s}^{\left(\text{RS}\right)}-\mathbf{y}^{\prime}\rVert^{2}. Then, the MSE is given by

𝔼[ε|𝐇^]=\displaystyle\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]= 2aci=1M{ϕ^(i,c)}2j=1Maj{ϕ^(j,j)}+2j=1Maj2(l=1M|ϕ^(l,j)|2+Mσe,i2𝐩j2)\displaystyle-2a_{c}\sum_{i=1}^{M}\Re{\left\{\hat{\phi}^{\left(i,c\right)}\right\}}-2\sum_{j=1}^{M}a_{j}\Re{\left\{\hat{\phi}^{\left(j,j\right)}\right\}}+2\sum_{j=1}^{M}a_{j}^{2}\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right)
+2ac2(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2)+2i=1M1q=i+1Mr=1Mar2{ϕ^(i,r)ϕ^(q,r)}\displaystyle+2a_{c}^{2}\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right)+2\sum_{i=1}^{M-1}\sum_{q=i+1}^{M}\sum_{r=1}^{M}a_{r}^{2}\Re\left\{\hat{\phi}^{\left(i,r\right)^{*}}\hat{\phi}^{\left(q,r\right)}\right\}
+l=1Mj=1jlMac2ϕ^(l,c)ϕ^(j,c)+M(1+2σn2)+1,\displaystyle+\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}^{2}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}+M\left(1+2\sigma_{n}^{2}\right)+1, (9)

where ϕ^(i,c)=𝐡^l,𝐩c\hat{\phi}^{\left(i,c\right)}=\hat{\mathbf{h}}_{l,*}\mathbf{p}_{c} and ϕ^(i,l)=𝐡^i,𝐩l\hat{\phi}^{\left(i,l\right)}=\hat{\mathbf{h}}_{i,*}\mathbf{p}_{l} for all i,l[1,M]i,l\in\left[1,M\right]. The proof to obtain the last equation can be found in appendix A. The partial derivatives of (9) with respect to ac{a}_{c} and ai{a}_{i} are expressed by

𝔼[ε|𝐇^]ac=\displaystyle\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{c}}= 2l=1Mj=1jlMacϕ^(l,c)ϕ^(j,c)2i=1M{ϕ^(i,c)}+4ac(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2),\displaystyle 2\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}-2\sum_{i=1}^{M}\Re{\left\{\hat{\phi}^{\left(i,c\right)}\right\}}+4a_{c}\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right), (10)
𝔼[ε|𝐇^]ai=\displaystyle\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{i}}= 4ail=1M1q=l+1M[ϕ^(l,i)ϕ^(q,i)]2{ϕ^(i,i)}+4ai(l=1M|ϕ^(l,i)|2+Mσe,i2𝐩i2).\displaystyle 4a_{i}\sum_{l=1}^{M-1}\sum_{q=l+1}^{M}\Re\left[\hat{\phi}^{\left(l,i\right)^{*}}\hat{\phi}^{\left(q,i\right)}\right]-2\Re{\left\{\hat{\phi}^{\left(i,i\right)}\right\}}+4a_{i}\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,i\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{i}\rVert^{2}\right). (11)

The partial derivatives given by (10) and (11) represent the gradient of the MSE with respect to the power allocation coefficients. With the obtained gradient we can employ a gradient descent algorithm, which is an iterative procedure that finds a local minimum of a differentiable function. The key idea is to take small steps in the opposite direction of the gradient, since this is the direction of the steepest descent. Remark that the objective function used is convex and has no local minimum. Then, the recursions of the proposed APA-R technique are given by

ac[t+1]\displaystyle a_{c}\left[t+1\right] =ac[t]μ𝔼[|ε|2|𝐇^T]ac,\displaystyle=a_{c}\left[t\right]-\mu\frac{\partial\mathbb{E}\left[\lvert\varepsilon\rvert^{2}|\hat{\mathbf{H}}^{\text{T}}\right]}{\partial a_{c}},
ai[t+1]\displaystyle a_{i}\left[t+1\right] =ai[t]μ𝔼[|ε|2|𝐇^T]ai,\displaystyle=a_{i}\left[t\right]-\mu\frac{\partial\mathbb{E}\left[\lvert\varepsilon\rvert^{2}|\hat{\mathbf{H}}^{\text{T}}\right]}{\partial a_{i}}, (12)

where the parameter μ\mu represents the learning rate of the adaptive algorithm. At each iteration, the power constraint is analyzed. Then, the coefficients are scaled with a power scaling factor by 𝐚(RS)[n]=β𝐚(RS)[n]\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]=\beta\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right], where β=1tr(diag(𝐚(RS)[n]𝐚(RS)[n]))\beta=\sqrt{\frac{1}{\textrm{tr}\left(\textrm{diag}\left(\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]\odot\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]\right)\right)}} to ensure that the power constraint is satisfied. Algorithm 1 summarizes the proposed APA-R algorithm.

1 given 𝐇^\hat{\mathbf{H}}, 𝐏\mathbf{P}, and μ\mu;
2 𝐚[1]=𝟎\mathbf{a}\left[1\right]=\mathbf{0};
3 for n=2n=2 to ItI_{t} do
4       𝔼[ε|𝐇^]ac=2l=1Mj=1jlMacϕ^(l,c)ϕ^(j,c)2i=1M{ϕ^(i,c)}+4ac(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2)\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{c}}=2\sum\limits_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}-2\sum\limits_{i=1}^{M}\Re{\left\{\hat{\phi}^{\left(i,c\right)}\right\}}+4a_{c}\left(\sum\limits_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right);
5       𝔼[ε|𝐇^]ai=4ail=1M1q=l+1M[ϕ^(l,i)ϕ^(q,i)]2{ϕ^(i,i)}+4ai(l=1M|ϕ^(l,i)|2+Mσe,i2𝐩i2)\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{i}}=4a_{i}\sum\limits_{l=1}^{M-1}\sum\limits_{q=l+1}^{M}\Re\left[\hat{\phi}^{\left(l,i\right)^{*}}\hat{\phi}^{\left(q,i\right)}\right]-2\Re{\left\{\hat{\phi}^{\left(i,i\right)}\right\}}+4a_{i}\left(\sum\limits_{l=1}^{M}\lvert\hat{\phi}^{\left(l,i\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{i}\rVert^{2}\right);
6       ac[n]=ac[n1]μ𝔼[ε|𝐇^]aca_{c}\left[n\right]=a_{c}\left[n-1\right]-\mu\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{c}};
7       ai[n]=ai[n1]μ𝔼[ε|𝐇^]aia_{i}\left[n\right]=a_{i}\left[n-1\right]-\mu\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{i}};
8       if tr(diag(𝐚(RS)[n]𝐚(RS)[n]))1\textrm{\rm tr}\left(\textrm{\rm diag}\left(\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]\odot\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]\right)\right)\neq 1 then
9             β=1tr(diag(𝐚(RS)[n]𝐚(RS)[n]))\beta=\sqrt{\frac{1}{\textrm{tr}\left(\textrm{diag}\left(\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]\odot\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]\right)\right)}};
10             𝐚(RS)[n]=β𝐚(RS)[n]\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right]=\beta\mathbf{a}^{\left(\rm{RS}\right)}\left[n\right];
11            
12       end if
13      
14 end for
Algorithm 1 Robust Adaptive Power allocation

III-B Adaptive Power Allocation

In this section, a simplified version of the proposed APA-R algorithm is derived. The main objective is to reduce the complexity of each recursion of the adaptive algorithm and avoid the load of computing the statistical parameters of the imperfect CSIT, while reaping the benefits of RS systems. The power allocation problem can be reformulated as the constrained optimization problem given by

min𝐚𝔼[𝐬(RS)𝐲2]s.t.tr(𝐏(RS)diag(𝐚(RS)𝐚(RS))𝐏(RS)H)=Etr,\begin{gathered}\min_{\mathbf{a}}\mathbb{E}\left[\lVert\mathbf{s}^{\left(\text{RS}\right)}-\mathbf{y}^{\prime}\rVert^{2}\right]\\ \text{s.t.}~{}~{}\text{tr}\left(\mathbf{P}^{\left(\text{RS}\right)}\text{diag}\left(\mathbf{a}^{\left(\text{RS}\right)}\odot\mathbf{a}^{\left(\text{RS}\right)}\right)\mathbf{P}^{\left(\text{RS}\right)^{H}}\right)=E_{tr},\end{gathered} (13)

In this case, the MSE is equivalent to

𝔼[ε]=\displaystyle\mathbb{E}\left[\varepsilon\right]= 2acj=1M{ϕ(j,c)}2l=1Mal{ϕ(l,l)}+2(l=1Mac2|ϕ(l,c)|2+i=1Mj=1Maj2|ϕ(i,j)|2)\displaystyle-2a_{c}\sum_{j=1}^{M}\Re\left\{\phi^{\left(j,c\right)}\right\}-2\sum_{l=1}^{M}a_{l}\Re\left\{\phi^{\left(l,l\right)}\right\}+2\left(\sum_{l=1}^{M}a_{c}^{2}\lvert\phi^{\left(l,c\right)}\rvert^{2}+\sum_{i=1}^{M}\sum_{j=1}^{M}a_{j}^{2}\lvert\phi^{\left(i,j\right)}\rvert^{2}\right)
+2i=1M1q=i+1Mr=1Mar2{ϕ(i,r)ϕ(q,r)}+l=1Mj=1jlMac2ϕ(l,c)ϕ(j,c)+M(1+2σn2)+1,\displaystyle+2\sum_{i=1}^{M-1}\sum_{q=i+1}^{M}\sum_{r=1}^{M}a_{r}^{2}\Re\left\{\phi^{\left(i,r\right)^{*}}\phi^{\left(q,r\right)}\right\}+\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}^{2}\phi^{\left(l,c\right)^{*}}\phi^{\left(j,c\right)}+M\left(1+2\sigma_{n}^{2}\right)+1, (14)

where we considered that ϕ(i,c)=𝐡l,𝐩c\phi^{\left(i,c\right)}=\mathbf{h}_{l,*}\mathbf{p}_{c} and ϕ(i,l)=𝐡i,𝐩l\phi^{\left(i,l\right)}=\mathbf{h}_{i,*}\mathbf{p}_{l} for all i,l[1,M]i,l\in\left[1,M\right]. The proof to obtain (14) can be found in appendix B. Taking the partial derivatives of (14) with respect to the coefficients of 𝐚(RS)\mathbf{a}^{\left(\text{RS}\right)} we arrive at

𝔼[ε]ac\displaystyle\frac{\partial\mathbb{E}\left[\varepsilon\right]}{\partial a_{c}} =4aci=1M|ϕ(i,c)|2+2acl=1Mj=1jlMϕ(l,c)ϕ(j,c)2q=1M[ϕ(q,c)],\displaystyle=4a_{c}\sum_{i=1}^{M}\lvert\phi^{\left(i,c\right)}\rvert^{2}+2a_{c}\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\phi^{\left(l,c\right)^{*}}\phi^{\left(j,c\right)}-2\sum_{q=1}^{M}\Re\left[\phi^{\left(q,c\right)}\right], (15)
𝔼[ε]ai\displaystyle\frac{\partial\mathbb{E}\left[\varepsilon\right]}{\partial a_{i}} =4aij=1M|ϕ(j,i)|2+4air=1M1q=r+1M[ϕ(r,i)ϕ(q,i)]2ai[ϕ(i,i)],\displaystyle=4a_{i}\sum_{j=1}^{M}\lvert\phi^{\left(j,i\right)}\rvert^{2}+4a_{i}\sum_{r=1}^{M-1}\sum_{q=r+1}^{M}\Re\left[\phi^{\left(r,i\right)^{*}}\phi^{\left(q,i\right)}\right]-2a_{i}\Re\left[\phi^{\left(i,i\right)}\right], (16)

The power allocation coefficients are adapted using (15) and (16) in the following recursions:

ac[t+1]\displaystyle a_{c}\left[t+1\right] =ac[t]μ𝔼[ε]ac\displaystyle=a_{c}\left[t\right]-\mu\frac{\partial\mathbb{E}\left[\varepsilon\right]}{\partial a_{c}}
ai[t+1]\displaystyle a_{i}\left[t+1\right] =ai[t]μ𝔼[ε]ai.\displaystyle=a_{i}\left[t\right]-\mu\frac{\partial\mathbb{E}\left[\varepsilon\right]}{\partial a_{i}}. (17)

III-C Sum-Rate Performance

In this section, we derive closed-form expressions to compute the sum-rate performance of the proposed algorithms. Specifically, we employ the ergodic sum-rate (ESR) as the main performance metric. Before the computation of the ESR we need to find the average power of the received signal, which is given by

𝔼[|yk|2]=ac2|𝐡kT𝐩c|2+i=1Kai2|𝐡kT𝐩i|2+σw2.\mathbb{E}\left[\lvert y_{k}\rvert^{2}\right]=a_{c}^{2}\lvert\mathbf{h}_{k}^{\textrm{T}}\mathbf{p}_{c}\rvert^{2}+\sum_{i=1}^{K}a_{i}^{2}\lvert\mathbf{h}_{k}^{\textrm{T}}\mathbf{p}_{i}\rvert^{2}+\sigma_{w}^{2}. (18)

It follows that the instantaneous SINR while decoding the common symbol is given by

γc,k\displaystyle\gamma_{c,k} =ac2|𝐡^kT𝐩c|2i=1Kai2|𝐡kT𝐩i|2+σw2.\displaystyle=\frac{a_{c}^{2}\lvert\mathbf{\hat{h}}_{k}^{\textrm{T}}\mathbf{p}_{c}\rvert^{2}}{\sum\limits_{i=1}^{K}a_{i}^{2}\lvert\mathbf{h}_{k}^{\textrm{T}}\mathbf{p}_{i}\rvert^{2}+\sigma_{w}^{2}}. (19)

Once the common symbol is decoded, we apply SIC to remove it from the received signal. Afterwards, we calculate the instantaneous SINR when decoding the kkth private stream, which is given by

γk=ak2|𝐡^kT𝐩k|2i=1ikKai2|𝐡k𝐩i|2+σw2.\gamma_{k}=\frac{a_{k}^{2}\lvert\mathbf{\hat{h}}_{k}^{\textrm{T}}\mathbf{p}_{k}\rvert^{2}}{\sum\limits_{\begin{subarray}{c}i=1\\ i\neq k\end{subarray}}^{K}a_{i}^{2}\lvert\mathbf{h}_{k}\mathbf{p}_{i}\rvert^{2}+\sigma_{w}^{2}}. (20)

Considering Gaussian signaling, the instantaneous common rate can be found with the following equation:

Rc,k=log2(1+γc,k).R_{c,k}=\log_{2}\left(1+\gamma_{c,k}\right). (21)

The private rate of the kkth stream is given by

Rk=log2(1+γk)R_{k}=\log_{2}\left(1+\gamma_{k}\right) (22)

Since imperfect CSIT is being considered, the instantaneous rates are not achievable. To that end, we employ the average sum rate (ASR) to average out the effect of the error in the channel estimate. The average common rate and the average private rate are given by

R¯c,k\displaystyle\bar{R}_{c,k} =𝔼[Rc,k|𝐆^]\displaystyle=\mathbb{E}\left[R_{c,k}|\mathbf{\hat{G}}\right] R¯k=𝔼[Rk|𝐆^],\displaystyle\bar{R}_{k}=\mathbb{E}\left[R_{k}|\mathbf{\hat{G}}\right], (23)

respectively. With the ASR we can obtain the ergodic sum-rate (ESR), which quantifies the performance of the system over a large number of channel realizations and is given by

Sr=mink𝔼[Rc,k¯]+k=1K𝔼[R¯k],S_{r}=\min_{k}\mathbb{E}\left[\bar{R_{c,k}}\right]+\sum_{k=1}^{K}\mathbb{E}\left[\bar{R}_{k}\right], (24)

IV Analysis

In this section, we carry out a convexity analysis and a statistical analysis of the proposed algorithms along with an assessment of their computational complexity in terms of floating point operations (FLOPs). Moreover, we derive a bound that establishes that the proposed APA-R algorithm is superior or comparable to the proposed APA algorithm.

IV-A Convexity analysis

In this section, we perform a convexity analysis of the optimization problem that gives rise to the proposed APA-R and APA algorithms. In order to establish convexity, we need to compute the second derivative of 𝔼[ε|𝐇^]\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right] with respect to aca_{c} and aia_{i}, and then check if it is greater than zero [47], i.e.,

2𝔼[ε|𝐇^]acac>0and2𝔼[ε|𝐇^]aiai>0,i=1,2,,K\frac{\partial^{2}\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{c}\partial a_{c}}>0~{}{\rm and}~{}\frac{\partial^{2}\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{i}\partial a_{i}}>0,~{}i=1,2,\ldots,K (25)

Let us first compute 2𝔼[ε|𝐇^]acac\frac{\partial^{2}\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{c}\partial a_{c}} using the results in (10):

ac(𝔼[ε|𝐇^]ac)=ac(2l=1Mj=1jlMacϕ^(l,c)ϕ^(j,c)2i=1M{ϕ^(i,c)}+4ac(i=1M|ϕ^(i,c)|2+Mσe,i2))=2l=1Mj=1jlMϕ^(l,c)ϕ^(j,c)+4(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2).\begin{split}\frac{\partial}{\partial a_{c}}\Bigg{(}\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{c}}\Bigg{)}&=\frac{\partial}{\partial a_{c}}\Bigg{(}2\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}-2\sum_{i=1}^{M}\Re{\left\{\hat{\phi}^{\left(i,c\right)}\right\}}+4a_{c}\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\right)\Bigg{)}\\ &=2\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}+4\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right).\end{split} (26)

Now let us compute 2𝔼[ε|𝐇^]aiai\frac{\partial^{2}\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{i}\partial a_{i}} using the results in (11):

ai(𝔼[ε|𝐇^]ai)=ai(4ail=1M1q=l+1M[ϕ^(l,i)ϕ^(q,i)]2{ϕ^(i,i)}+4ai(l=1M|ϕ^(l,i)|2+Mσe,i2𝐩i2))=4l=1M1q=l+1M[ϕ^(l,i)ϕ^(q,i)]+4(l=1M|ϕ^(l,i)|2+Mσe,i2𝐩i2),\begin{split}\frac{\partial}{\partial a_{i}}\Bigg{(}\frac{\partial\mathbb{E}\left[\varepsilon|\hat{\mathbf{H}}\right]}{\partial a_{i}}\Bigg{)}&=\frac{\partial}{\partial a_{i}}\Bigg{(}4a_{i}\sum_{l=1}^{M-1}\sum_{q=l+1}^{M}\Re\left[\hat{\phi}^{\left(l,i\right)^{*}}\hat{\phi}^{\left(q,i\right)}\right]-2\Re{\left\{\hat{\phi}^{\left(i,i\right)}\right\}}+4a_{i}\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,i\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{i}\rVert^{2}\right)\Bigg{)}\\ &=4\sum_{l=1}^{M-1}\sum_{q=l+1}^{M}\Re\left[\hat{\phi}^{\left(l,i\right)^{*}}\hat{\phi}^{\left(q,i\right)}\right]+4\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,i\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{i}\rVert^{2}\right),\end{split} (27)

Since we have the sum of the strictly convex terms in (26) and (27) the objective function associated with APA-R is strictly convex [47]. The power constraint is also strictly convex and only scales the powers to be adjusted. In the case of the APA algorithm, the objective function does not employ knowledge of the error variance σe,i2\sigma_{e,i}^{2} and remains strictly convex.

IV-B Bound on the MSE of APA and APA-R

Let us now show that the proposed APA-R power allocation produces a lower MSE than that of the proposed APA power allocation. The MSE obtained in (14) assumes that the transmitter has perfect knowledge of the channel. Under such assumption the optimal coefficients 𝐚o\mathbf{a}_{o} that minimize the error are found. However, under imperfect CSIT the transmitter is unaware of 𝐇~\tilde{\mathbf{H}} and the adaptive algorithm performs the power allocation by employing 𝐇^\hat{\mathbf{H}} instead of 𝐇\mathbf{H}. This results in a power allocation given by 𝐚^(APA)\hat{\mathbf{a}}^{\left(\text{APA}\right)} which originates an increase in the MSE obtained. It follows that

ε(𝐇,𝐚o)ε(𝐇,𝐚^(APA))\varepsilon\left(\mathbf{H},\mathbf{a}_{o}\right)\leq\varepsilon\left(\mathbf{H},\hat{\mathbf{a}}^{\left(\text{APA}\right)}\right) (28)

On the other hand, the robust adaptive algorithm finds the optimal 𝐚o(APA-R)\mathbf{a}_{o}^{\left(\text{APA-R}\right)} that minimizes 𝔼[ε(𝐇,𝐚)|𝐇^]\mathbb{E}\left[\varepsilon\left(\mathbf{H},\mathbf{a}\right)\right|\hat{\mathbf{H}}] and therefore takes into account that only partial knowledge of the channel is available. Since the coefficients 𝐚^(APA)\hat{\mathbf{a}}^{\left(\text{APA}\right)} and 𝐚o(APA-R)\mathbf{a}_{o}^{\left(\text{APA-R}\right)} are different, we have

𝔼[ε(𝐇,𝐚𝐨(APA-R))|𝐇^]𝔼[ε(𝐇,𝐚^(APA))|𝐇^]\mathbb{E}\left[\varepsilon\left(\mathbf{H,\mathbf{a}_{o}^{\left(\text{APA-R}\right)}}\right)|\hat{\mathbf{H}}\right]\leq\mathbb{E}\left[\varepsilon\left(\mathbf{H},\hat{\mathbf{a}}^{\left(\text{APA}\right)}\right)|\hat{\mathbf{H}}\right] (29)

Note that under perfect CSIT equation (9) reduces to (14). In such circumstances 𝐚o(APA)=𝐚o(APA-R)\mathbf{a}_{o}^{\left(\text{APA}\right)}=\mathbf{a}_{o}^{\left(\text{APA-R}\right)} and therefore both algorithms are equivalent. In the following, we evaluate the performance obtained by the proposed algorithms. Specifically, we have that 𝐚^(APA)=𝐚o(APA-R)+𝐚e\hat{\mathbf{a}}^{\left(\text{APA}\right)}=\mathbf{a}_{o}^{\left(\text{APA-R}\right)}+\mathbf{a}_{e}, where 𝐚e=[ac,e,a1,e,,aM,e]T\mathbf{a}_{e}=\left[a_{c,e},a_{1,e},\cdots,a_{M,e}\right]^{\text{T}} is the error produced from the assumption that the BS has perfect CSIT. Then, we have

𝔼[ε(APA)ε](APA-R)\displaystyle\mathbb{E}\left[\varepsilon^{\left(\text{APA}\right)}-\varepsilon\right.\left.{}^{\left(\text{APA-R}\right)}\right] =2ac,ei=1M{ϕ^(i,c)}2j=1Maj,e{ϕ^(j,j)}+l=1Mj=1jlMac,e2ϕ^(l,c)ϕ^(j,c)\displaystyle=-2a_{c,e}\sum_{i=1}^{M}\Re{\left\{\hat{\phi}^{\left(i,c\right)}\right\}}-2\sum_{j=1}^{M}a_{j,e}\Re{\left\{\hat{\phi}^{\left(j,j\right)}\right\}}+\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c,e}^{2}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}
+2ac,e2(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2)+2i=1M1q=i+1Mr=1Mar,e2[ϕ^(i,r)ϕ^(q,r)]\displaystyle+2a_{c,e}^{2}\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right)+2\sum_{i=1}^{M-1}\sum_{q=i+1}^{M}\sum_{r=1}^{M}a_{r,e}^{2}\Re\left[\hat{\phi}^{\left(i,r\right)^{*}}\hat{\phi}^{\left(q,r\right)}\right]
+2j=1Maj,e2(l=1M|ϕ^(l,j)|2+Mσe,i2𝐩j2).\displaystyle+2\sum_{j=1}^{M}a_{j,e}^{2}\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right). (30)

which is a positive quantity when 2ac,ei=1M{ϕ^(i,c)}<2ac,e2(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2)-2a_{c,e}\sum_{i=1}^{M}\Re{\left\{\hat{\phi}^{\left(i,c\right)}\right\}}<2a_{c,e}^{2}\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right) and 2j=1Maj,e{ϕ^(j,j)}<2j=1Maj,e2(l=1M|ϕ^(l,j)|2+Mσe,i2𝐩j2)-2\sum_{j=1}^{M}a_{j,e}\Re{\left\{\hat{\phi}^{\left(j,j\right)}\right\}}<2\sum_{j=1}^{M}a_{j,e}^{2}\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right). The inequalities hold as long as ac,e[i=1M({ϕ^(i,c)})2+i=1M({ϕ^(i,c)})2+Mσe,i2𝐩c2]>i=1M{ϕ(i,c)}a_{c,e}\left[\sum_{i=1}^{M}\left(\Re\left\{\hat{\phi}^{\left(i,c\right)}\right\}\right)^{2}+\sum_{i=1}^{M}\left(\Im\left\{\hat{\phi}^{\left(i,c\right)}\right\}\right)^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right]>\sum_{i=1}^{M}\Re{\left\{\phi^{\left(i,c\right)}\right\}} and j=1Maj,e[l=1M({ϕ^(l,j)})2+l=1M({ϕ^(l,j)})2+Mσe,i2𝐩j2]>j=1M{ϕ(j,j)}\sum_{j=1}^{M}a_{j,e}\left[\sum_{l=1}^{M}\left(\Re\left\{\hat{\phi}^{\left(l,j\right)}\right\}\right)^{2}+\sum_{l=1}^{M}\left(\Im\left\{\hat{\phi}^{\left(l,j\right)}\right\}\right)^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right]>\sum_{j=1}^{M}\Re{\left\{\phi^{\left(j,j\right)}\right\}}. As the error in the power allocation coefficients grows the left-hand side of the two last inequalities increases. This explains why the proposed APA-R performs better than the proposed APA algorithm.

IV-C Statistical Analysis

The performance of adaptive learning algorithms is usually measured in terms of its transient behavior and steady-state behaviour. These measurements provide information about the stability, the convergence rate, and the MSE achieved by the algorithm[53, 54]. Let us consider the adaptive power allocation with the update equations given by (17). Expanding the terms of (17) for the private streams, we get

aj[t+1]=\displaystyle a_{j}\left[t+1\right]= aj[t]4μaj[t]l=1M|ϕ(l,j)|2+2μ{ϕ(j,j)}4μaj[t]q=1M1r=q+1M{ϕ(q,j)ϕ(r,j)}.\displaystyle a_{j}\left[t\right]-4\mu a_{j}\left[t\right]\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+2\mu\Re\left\{\phi^{\left(j,j\right)}\right\}-4\mu a_{j}\left[t\right]\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}\Re\left\{\phi^{\left(q,j\right)^{*}}\phi^{\left(r,j\right)}\right\}. (31)

Let us define the error between the estimate of the power coefficients and the optimal parameters as follows:

eaj[t]=aj[t]aj(o),e_{a_{j}}\left[t\right]=a_{j}\left[t\right]-a_{j}^{\left(o\right)}, (32)

where aj(o)a_{j}^{\left(o\right)} represents the optimal allocation for the jjth coefficient.

By subtracting (32) from (31), we obtain

eaj[t+1]=\displaystyle e_{a_{j}}\left[t+1\right]= eaj[t]4μaj[t]l=1M|ϕ(l,j)|2+2μ{ϕ(j,j)}4μaj[t]q=1M1r=q+1M{ϕ(q,j)ϕ(r,j)}\displaystyle e_{a_{j}}\left[t\right]-4\mu a_{j}\left[t\right]\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+2\mu\Re\left\{\phi^{\left(j,j\right)}\right\}-4\mu a_{j}\left[t\right]\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}\Re\left\{\phi^{\left(q,j\right)^{*}}\phi^{\left(r,j\right)}\right\}
=\displaystyle= eaj[t]+2μ{ϕ(j,j)}4μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))aj[t],\displaystyle e_{a_{j}}\left[t\right]+2\mu\Re\left\{\phi^{\left(j,j\right)}\right\}-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)a_{j}\left[t\right], (33)

where fq,r(j)={(𝐡q,𝐩j)(𝐡r,𝐩j)}f_{q,r}^{\left(j\right)}=\Re\left\{\left(\mathbf{h}_{q,*}\mathbf{p}_{j}\right)^{*}\left(\mathbf{h}_{r,*}\mathbf{p}_{j}\right)\right\}. Expanding the terms in (33), we get

eaj[t+1]=\displaystyle e_{a_{j}}\left[t+1\right]= 4μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))aj(o)+eaj[t]+2μ{ϕ(j,j)}\displaystyle-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)a_{j}^{\left(o\right)}+e_{a_{j}}\left[t\right]+2\mu\Re\left\{\phi^{\left(j,j\right)}\right\}
4μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))eaj[t].\displaystyle-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)e_{a_{j}}\left[t\right].

Rearranging the terms of the last equation, we obtain

eaj[t+1]\displaystyle e_{a_{j}}\left[t+\right.\left.1\right] ={14μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))}eaj[t]\displaystyle=\left\{1-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)\right\}e_{a_{j}}\left[t\right]
4μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))aj(o)+2μ{ϕ(j,j)}.\displaystyle-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)a_{j}^{\left(o\right)}+2\mu\Re\left\{\phi^{\left(j,j\right)}\right\}. (34)

Equation (34) can be rewritten as follows

eaj[t+1]=\displaystyle e_{a_{j}}\left[t+\right.\left.1\right]= {14μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))}eaj[t]\displaystyle\left\{1-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)\right\}e_{a_{j}}\left[t\right]
+2μ(MSEmin(aj(o))aj(o)aj(o)q=1M1r=q+1Mfq,r(j)),\displaystyle+2\mu\left(\frac{\text{MSE}_{\textrm{min}}\left(a_{j}^{\left(o\right)}\right)}{a_{j}^{\left(o\right)}}-a_{j}^{\left(o\right)}\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right), (35)

where

MSEmin(aj(o))=2aj(o)(aj(o)i=1M|ϕ(i,j)|2{ϕ(j,j)}+q=1M1r=q+1Mfq,r(j)).\displaystyle\text{MSE}_{\rm min}\left(a_{j}^{\left(o\right)}\right)=2a_{j}^{\left(o\right)}\left(a_{j}^{\left(o\right)}\sum_{i=1}^{M}\lvert\phi^{\left(i,j\right)}\rvert^{2}-\Re\left\{\phi^{\left(j,j\right)}\right\}\right.\left.+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right). (36)

Bu multiplying (35) by ea[t+1]e_{a}\left[t+1\right] and taking the expected value, we obtain

σeaj2[t1+]=\displaystyle\sigma_{e_{a_{j}}}^{2}\left[t\right.\left.1+\right]= {14μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))}2σeaj2[t]\displaystyle\left\{1-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)\right\}^{2}\sigma_{e_{a_{j}}}^{2}\left[t\right]
+4μ2(MSEmin(aj(o))aj(o)aj(o)q=1M1r=q+1Mfq,r(j))2,\displaystyle+4\mu^{2}\left(\frac{\text{MSE}_{min}\left(a_{j}^{\left(o\right)}\right)}{a_{j}^{\left(o\right)}}-a_{j}^{\left(o\right)}\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)^{2}, (37)

where we consider that 𝔼[eaj[i]]𝟎\mathbb{E}\left[e_{a_{j}}\left[i\right]\right]\approx\mathbf{0}.The previous equation constitutes a geometric series with geometric ratio equal to 14μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))1-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right). Then, we have

|14μ(l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j))|<1\left\lvert 1-4\mu\left(\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)\right\rvert<1 (38)

Note that from the last equation the step size must fulfill

0<μj<12λj,0<\mu_{j}<\frac{1}{2\lambda_{j}}, (39)

with λj=l=1M|ϕ(l,j)|2+q=1M1r=q+1Mfq,r(j)\lambda_{j}=\sum_{l=1}^{M}\lvert\phi^{\left(l,j\right)}\rvert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}.

For the common power allocation coefficient we have the following recursion:

ac[t+1]=\displaystyle a_{c}\left[t+1\right]= ac[t]4μac[t]j=1M|ϕ(j,c)|22μl=1M{ϕ(l,c)}+2μac[t]f(c),\displaystyle a_{c}\left[t\right]-4\mu a_{c}\left[t\right]\sum_{j=1}^{M}\lvert\phi^{\left(j,c\right)}\rvert^{2}-2\mu\sum_{l=1}^{M}\Re\left\{\phi^{\left(l,c\right)}\right\}+2\mu a_{c}\left[t\right]f^{\left(c\right)}, (40)

where f(c)=q=1Mr=1rqMϕ(q,c)ϕ(r,c)f^{\left(c\right)}=\sum_{q=1}^{M}\sum_{\begin{subarray}{c}r=1\\ r\neq q\end{subarray}}^{M}\phi^{\left(q,c\right)^{*}}\phi^{\left(r,c\right)}. The error with respect to the optimal power allocation of the common stream is given by

ec[i]=ac[i]ac(o).e_{c}\left[i\right]=a_{c}{\left[i\right]}-a_{c}^{\left(o\right)}. (41)

Following a similar procedure to the one employed for the private streams we arrive at

ec[t+1]=\displaystyle e_{c}\left[t+1\right]= {14μj=1M|ϕ(j,c)|22μf(c)}ec[t]2μ(2j=1M|ϕ(j,c)|2+μf(c))ac(o)\displaystyle\left\{1-4\mu\sum_{j=1}^{M}\lvert\phi^{\left(j,c\right)}\rvert^{2}-2\mu f^{\left(c\right)}\vphantom{\sum_{j=1}^{M}}\right\}e_{c}\left[t\right]-2\mu\left(2\sum_{j=1}^{M}\lvert\phi^{\left(j,c\right)}\rvert^{2}+\mu f^{\left(c\right)}\right)a_{c}^{\left(o\right)}
2μlM{ϕ(l,c)}.\displaystyle-2\mu\sum_{l}^{M}\Re\left\{\phi^{\left(l,c\right)}\right\}. (42)

Multiplying the previous equation by ec[t+1]e_{c}\left[t+1\right] and taking the expected value leads us to:

σec2[t+1]=\displaystyle\sigma_{e_{c}}^{2}\left[t+1\right]= {12μ(2j=1M|ϕ(j,c)|2f(c))}σec2[t]\displaystyle\left\{1-2\mu\left(2\sum_{j=1}^{M}\lvert\phi^{\left(j,c\right)}\rvert^{2}-f^{\left(c\right)}\right)\right\}\sigma_{e_{c}}^{2}\left[t\right]
4(MSEmin(ac(o))ac(o)+ac(o)j=1M{ϕ(j,c)})\displaystyle-4\left(\frac{\text{MSE}_{\text{min}}\left(a_{c}^{\left(o\right)}\right)}{a_{c}^{\left(o\right)}}+a_{c}^{\left(o\right)}\sum_{j=1}^{M}\Re\left\{\phi^{\left(j,c\right)}\right\}\right) (43)

It follows that the geometric ratio of the recursion is equal to 12μ(2j=1M|ϕ(j,c)|2f(c))1-2\mu\left(2\sum\limits_{j=1}^{M}\lvert\phi^{\left(j,c\right)}\rvert^{2}-f^{\left(c\right)}\right). Then, the step size must be in the following range:

0<μc<1λc,0<\mu_{c}<\frac{1}{\lambda_{c}}, (44)

where

λc=2j=1M|ϕ(j,c)|2f(c)\lambda_{c}=2\sum_{j=1}^{M}\lvert\phi^{\left(j,c\right)}\rvert^{2}-f^{\left(c\right)} (45)

The step-size of the algorithm must be less or equal to min(μc,μj)\min\left(\mu_{c},\mu_{j}\right) j{1,2,,M}\forall j\in\left\{1,2,\cdots,M\right\}. The stability bounds provide useful information on the choice of the step sizes.

Let us now consider the APA-R algorithm. The a posteriori error can be expressed as follows:

eaj[t+1]=\displaystyle e_{a_{j}}\left[t+1\right]= eaj[t]+2μ{ϕ^(j,j)}4μ(l=1M|ϕ^(l,j)|2+Mσe,i2𝐩j2+q=1M1r=q+1Mfq,r(j))aj[t].\displaystyle e_{a_{j}}\left[t\right]+2\mu\Re{\left\{\hat{\phi}^{\left(j,j\right)}\right\}}-4\mu\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right.\left.+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)a_{j}\left[t\right]. (46)

Expanding the terms of the last equation, we get

eaj[t+1]=\displaystyle e_{a_{j}}\left[t+1\right]= {14μ(l=1M|ϕ^(l,j)|2+Mσe,i2𝐩j2+q=1M1r=q+1Mfq,r(j))}eaj[t]\displaystyle\left\{1-4\mu\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right.\right.\left.\left.+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)\right\}e_{a_{j}}\left[t\right]
4μ(l=1M|ϕ^(l,j)|2+Mσe,i2𝐩j2+q=1M1r=q+1Mfq,r(j))aj(o)+2μ{ϕ^(j,j)}.\displaystyle-4\mu\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right.\left.+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)a_{j}^{\left(o\right)}+2\mu\Re\left\{\hat{\phi}^{\left(j,j\right)}\right\}. (47)

The geometric ratio of the robust APA algorithm is given by 14μ(l=1M|ϕ^(l,j)|2+σe,i2𝐩j21-4\mu\left(\sum\limits_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right. +q=1M1r=q+1Mfq,r(j))\left.+\sum\limits_{q=1}^{M-1}\sum\limits_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right). Then, we have that

|14μ(l=1M|ϕ^(l,j)|2+σe,i2𝐩j2+q=1M1r=q+1Mfq,r(j))|<1\left\lvert 1-4\mu\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}\right)\right\rvert<1 (48)

Therefore, the step size must satisfy the following inequality:

0<μj(r)<12λj(r),0<\mu_{j}^{\left(\textrm{r}\right)}<\frac{1}{2\lambda^{\left(\textrm{r}\right)}_{j}}, (49)

where λj(r)=l=1M|ϕ^(l,j)|2+σe,i2𝐩j2+q=1M1r=q+1Mfq,r(j)\lambda^{\left(\textrm{r}\right)}_{j}=\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}+\sum_{q=1}^{M-1}\sum_{r=q+1}^{M}f_{q,r}^{\left(j\right)}. Following a similar procedure for the power coefficient of the common stream lead us to

0<μc(r)<1λc(r),0<\mu_{c}^{\left(\textrm{r}\right)}<\frac{1}{\lambda^{\left(\textrm{r}\right)}_{c}}, (50)

where

λc(r)=2j=1M|ϕ^(j,c)|2+σe,i2𝐩j2f(c)\lambda_{c}^{\left(\text{r}\right)}=2\sum_{j=1}^{M}\lvert\hat{\phi}^{\left(j,c\right)}\rvert^{2}+\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}-f^{\left(c\right)} (51)

As in the previous case, the step size is chosen using min(μc(r),μj(r))\min\left(\mu^{\left(\textrm{r}\right)}_{c},\mu^{\left(\textrm{r}\right)}_{j}\right) j{1,2,,M}\forall j\in\left\{1,2,\cdots,M\right\}. The variable λc(r)\lambda_{c}^{\left(\textrm{r}\right)} has an additional term given by σe,i2𝐩j2\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2} when compared to λc\lambda_{c} of the APA algorithm. In this sense, the upper bound of (70) is smaller than the bound in (65). In other words, the step size of APA-R takes smaller values than the step size of APA.

IV-D Computational Complexity

In this section the number of FLOPs performed by the proposed algorithms is computed. For this purpose let us consider the following results to simplify the analysis. Consider the complex vector 𝐳1\mathbf{z}_{1} and 𝐳2n\mathbf{z}_{2}\in\mathbb{C}^{n}. Then, we have the following results:

  • The product 𝐳1T𝐳2\mathbf{z}_{1}^{\text{T}}\mathbf{z}_{2} requires 8n28n-2 FLOPs.

  • The term 𝐳12\lVert\mathbf{z}_{1}\rVert^{2} requires 7n27n-2 FLOPs

The gradient in equation (16) requires the computation of three terms. The first term, which is given by 4aci=1M|𝐡i,𝐩c|24a_{c}\sum\limits_{i=1}^{M}\lvert\mathbf{h}_{i,*}\mathbf{p}_{c}\rvert^{2} needs a total of 8NtM+2M+18N_{t}M+2M+1 FLOPs. The evaluation of the second term results in 8NtM+M8N_{t}M+M. For the last term we have a total of (9M29M+2)/2\left(9M^{2}-9M+2\right)/2. Considering a system where Nt=Nr=M=nN_{t}=N_{r}=M=n we have that the number of FLOPs required by the proposed adaptive algorithm is given by 412n3+19n2+52n+4\frac{41}{2}n^{3}+19n^{2}+\frac{5}{2}n+4.

The computational complexity of the gradients in (10) and (11) can be computed in a similar manner. However, in this case we have an additional term given by 4acMσe,i2𝐩i24a_{c}M\sigma_{e,i}^{2}\lVert\mathbf{p}_{i}\rVert^{2}, which requires a total of 7Nt+27N_{t}+2 FLOPs. Then, the robust adaptive algorithm requires a total of 412n3+19n2+192n+6\frac{41}{2}n^{3}+19n^{2}+\frac{19}{2}n+6. It is important to mention that the computational complexity presented above represents the number of FLOPs of the adaptive algorithm per iteration.

In contrast, the optimal power allocation for the conventional SDMA system requires the exhaustive search over 𝐀\mathbf{A} with a fine grid. Given a system with 1212 streams and a grid step of 0.0010.001, the exhaustive search would require the evaluation of 50050005005000 different power allocation matrices for each channel realization. In contrast, the adaptive approaches presented require only the computation of around 3030 iterations. Furthermore, the complexity of the exhaustive search for an RS system is even higher since the search is perform over 𝐀(RS)\mathbf{A}^{\left(\text{RS}\right)}, which additionally contains the power allocated to the common stream.

Table LABEL:Computational_complexity_power_allocation summarizes the computational complexity of the proposed algorithms employing the big 𝒪\mathcal{O} notation. In Table LABEL:Computational_complexity_power_allocation, IoI_{o} denotes the number of points of the grid given a step size, IwI_{w} refers to the number of iterations of the alternating procedure and IaI_{a} denotes the number of iterations for the adaptive algorithms. It is worth noting that Io>>IaI_{o}>>I_{a}. Moreover, the inner iterations employed in the WMMSE approach are much more demanding than the iterations of the proposed APA and APA-R algorithms. Fig 3(a) shows the computational complexity in terms of FLOPS assuming that the number of transmit antennas increases. The term CF represents the closed-form power allocation in [45], which requires the inversion of an Nt×NtN_{t}\times N_{t} matrix. The step of the grid was set to 0.010.01 for the ES and the number of iterations to 3030 for the WMMSE, APA and APA-R approaches. Note that in practice the WMMSE iterates continuously until meeting a predefined accuracy. In general, it requires more than 3030 iterations. It is also important to point out that the cost per iteration of the adaptive approaches can be reduced after the first iteration. The precoders and the channel are fixed given a transmission block. Therefore, after the first iteration, we can avoid the multiplication of the precoder by the channel matrix in the update equation. In contrast, the WMMSE must perform the whole procedure again since the precoders are updated between iterations. This is illustrated in Fig. 3(b).

TABLE I: Computational complexity of the power allocation algorithms.
Technique 𝒪\mathcal{O}
SDMA-ES 𝒪(NtIo2M3)\mathcal{O}\left(N_{t}I_{o}^{2}M^{3}\right)
WMMSE 𝒪(IwNtM3)\mathcal{O}\left(I_{w}N_{t}M^{3}\right)
RS-ES 𝒪(NtIo2(M+1)3)\mathcal{O}\left(N_{t}I_{o}^{2}(M+1)^{3}\right)
RS-APA 𝒪(IaNt(M+1)2)\mathcal{O}\left(I_{a}N_{t}(M+1)^{2}\right)
RS-APA-R 𝒪(IaNt(M+1)2)\mathcal{O}\left(I_{a}N_{t}(M+1)^{2}\right)
CF[45] 𝒪(Nt3)\mathcal{O}\left(N_{t}^{3}\right)
Refer to caption
(a) Number of FLOPs required by different power allocation algorithms considering a MU-MIMO system with an increasing number of antennas.

. Refer to caption

(b) Number of FLOPs required per iteration considering a MU-MIMO system with Nt=16Nt=16 and M=16M=16.
Figure 3: Computational Complexity

V Simulations

In this section, the performance of the proposed APA-R and APA algorithms is assessed against existing power allocation approaches, namely, the ES, the WMMSE [42], the closed-form approach of [45], the power allocation obtained directly from the precoders definition which is denoted here as random power allocation, and the uniform power allocation (UPA) approaches. Unless otherwise specified, we consider an RS MU-MIMO system where the BS is equipped with four antennas and transmits data to two users, each one equipped with two antennas. The inputs are statistically independent and follow a Gaussian distribution. A flat fading Rayleigh channel, which remains fixed during the transmission of a packet, is considered, we assume additive white Gaussian noise with zero mean and unit variance, and the SNR varies with EtrE_{tr}. For all the experiments, the common precoder is computed by employing a SVD over the channel matrix, i.e. 𝐇^=𝐒𝚿𝐕H\hat{\mathbf{H}}=\mathbf{S}\mathbf{\Psi}\mathbf{V}^{H}. Then we set the common precoder equal to the first column of matrix 𝐕\mathbf{V}, i.e., 𝐩c=𝐯1\mathbf{p}_{c}=\mathbf{v}_{1}.

In the first example, we consider the transmission under imperfect CSIT. The proposed APA-R algorithm is compared against the closed form expression from [45] and against ES with random and UPA algorithms. The first ES approach fixes a random power allocation for the private streams and then an exhaustive search is carried out to find the best power allocation for the common stream. The second scheme considers that the power is uniformly distributed among private streams and then performs an exhaustive search to find the optimum value for aca_{c}. Fig. 4 shows the performance obtained with a MF. Although ES obtains the best performance, it requires a great amount of resources in terms of computational complexity. Moreover, we can see that the closed-form power allocation does not allocate power to the common message in the low SNR regime. The reason for this behavior is that this power allocation scheme was derived for massive MIMO environments where the excess of transmit antennas gets rid of the residual interference at low SNR and no common message is required. As the SNR increases the residual interference becomes larger and the algorithm allocates power to the common stream.

Refer to caption
Figure 4: Sum-rate performance of RS-MF precoding scheme, Nt=4N_{t}=4, Nk=4N_{k}=4, K=1K=1, and σe2=0.05\sigma_{e}^{2}=0.05.

In the next example, the ZF precoder has been considered. The results illustrated in Fig. 5 show that the techniques that perform ES, which are termed as RS-ZF-ES+Random and RS-ZF-ES-UPA, have the best performance. However, the APA and APA-R adaptive algorithms obtain a consistent gain when compared to the conventional ZF algorithm.

Refer to caption
Figure 5: Sum-rate performance of RS-ZF precoding scheme, Nt=4N_{t}=4, Nk=2N_{k}=2, K=2K=2, and σe2=0.1\sigma_{e}^{2}=0.1.

In Fig. 6 we employed the MMSE precoder. In this case, an exhaustive search was performed to obtain the best power allocation coefficients for all streams. The technique from [27] was also considered and is denoted in the figure as RS-WMMSE. We can observe that the best performance is obtained by ES. The proposed APA and APA-R algorithms obtain a consistent gain when compared to the conventional MMSE precoder. The performance is worse than that of ES and the RS-WMMSE, but the computational complexity is also much lower.

Refer to caption
Figure 6: Sum-rate performance of RS-MMSE precoding scheme, Nt=4N_{t}=4, Nk=2N_{k}=2, K=2K=2, and σe2=0.2\sigma_{e}^{2}=0.2.

In the next example, we evaluate the performance of the proposed APA and APA-R techniques as the error in the channel estimate becomes larger. For this scenario, we consider a fixed SNR equal to 20 dB. Fig. 7 depicts the results of different power allocation techniques. The results show that the APA-R performs better than the APA as the variance of the error increases.

Refer to caption
Figure 7: Sum-rate performance of RS-ZF precoding scheme, Nt=4N_{t}=4, Nk=2N_{k}=2, and K=2K=2.

Let us now consider the ESR obtained versus the number of iterations, which is shown in Fig. 8. The step size of the adaptive algorithms was set to 0.0040.004 and the SNR to 1010 dB. Fig. 8 shows that APA and APA-R obtain better performance than WMMSE with few iterations, i.e., with reduced cost. Recall that the cost per iteration is much lower for the adaptive algorithms.

Refer to caption
Figure 8: Sum-rate performance of RS-ZF precoding scheme, Nt=4N_{t}=4, Nk=1N_{k}=1,K=4K=4, and σe2=0.2\sigma_{e}^{2}=0.2.

In Fig. 9 we can notice the power allocated to the common stream. For this simulation we consider the same setup as in the previous simulation. We can observe that the parameter aca_{c} increases with the SNR. In other words, as the MUI becomes more significant, more power is allocated to the common stream. We can also notice that the proposed APA-R algorithm allocates more power to the common stream than that of the APA algorithm.

Refer to caption
Figure 9: Power allocated to the common stream, Nt=4N_{t}=4, Nk=2N_{k}=2, K=2K=2.

In the last example, we consider the ZF precoder in a MU-MIMO system where the BS is equipped with 2424 transmit antennas. The information is sent to 2424 users which are randomly distributed over the area of interest. Fig. 10 shows the results obtained by employing the proposed APA and APA-R algorithms. Specifically, it can be noticed that the RS system equipped with APA-R can obtain a gain of up to 20%20\% over that random allocation and up to 50%50\% over that of a conventional MU-MIMO system with random allocation.

Refer to caption
Figure 10: Sum-rate performance of RS-ZF precoding scheme, Nt=24N_{t}=24, Nk=1N_{k}=1, K=24K=24, and σe2=0.1\sigma_{e}^{2}=0.1.

VI Conclusion

In this work, adaptive power allocation techniques have been developed for RS-MIMO and conventional MU-MIMO systems. Differently to optimal and WMMSE power allocation often employed for RS-based systems that are computationally very costly, the proposed APA and APA-R algorithms have low computational complexity and require fewer iterations for new transmission blocks, being suitable for practical systems. Numerical results have shown that the proposed power allocation algorithms, namely APA and APA-R, are not very far from the performance of exhaustive search with uniform power allocation. Furthermore, the proposed robust technique, i.e., APA-R, increases the robustness of the system against CSIT imperfections.

Appendix A Proof of the MSE for the APA-R

In what follows the derivation of the MSE for the APA-R algorithm is detailed. Let us first expand the MSE, which is given by

𝔼[ε|𝐇^]=\displaystyle\mathbb{E}\left[\varepsilon\lvert\hat{\mathbf{H}}\right]= 𝔼[(𝐬(RS)𝐲)H(𝐬(RS)𝐲)|𝐇^]\displaystyle\mathbb{E}\left[\left(\mathbf{s}^{\left(\text{RS}\right)}-\mathbf{y}^{\prime}\right)^{H}\left(\mathbf{s}^{\left(\text{RS}\right)}-\mathbf{y}^{\prime}\right)\lvert\hat{\mathbf{H}}\right]
=\displaystyle= 𝔼[𝐬(RS)H𝐬(RS)|𝐇^]T1𝔼[𝐬(RS)H𝐲|𝐇^]T2𝔼[𝐲H𝐬(RS)|𝐇^]T3+𝔼[𝐲H𝐲|𝐇^]T4.\displaystyle\underbrace{\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{s}^{\left(\text{RS}\right)}\lvert\hat{\mathbf{H}}\right]}_{T_{1}}-\underbrace{\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{y}^{\prime}\lvert\hat{\mathbf{H}}\right]}_{T_{2}}-\underbrace{\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{s}^{\left(\text{RS}\right)}\lvert\hat{\mathbf{H}}\right]}_{T_{3}}+\underbrace{\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{y}^{\prime}\lvert\hat{\mathbf{H}}\right]}_{T_{4}}. (52)

The first term of (52) is independent from 𝐇^\hat{\mathbf{H}} and can be reduced to the following expression:

𝔼[𝐬(RS)H𝐬(RS)]=\displaystyle\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{s}^{\left(\text{RS}\right)}\right]= 𝔼[scsc]+𝔼[s1s1]+𝔼[sMsM],=M+1.\displaystyle\mathbb{E}\left[s_{c}^{*}s_{c}\right]+\mathbb{E}\left[s_{1}^{*}s_{1}\right]\cdots+\mathbb{E}\left[s_{M}^{*}s_{M}\right],=M+1. (53)

The second term given by T2T_{2} requires the computation of the following term:

𝐬(RS)H𝐲=\displaystyle\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{y}^{\prime}= scyc+s1y1++sMyM,\displaystyle s_{c}^{*}y_{c}+s_{1}^{*}y_{1}+\cdots+s_{M}^{*}y_{M},
=\displaystyle= scl=1M[acsc(𝐡^l,+𝐡~l,)𝐩c+j=1Majsj(𝐡^l,+𝐡~l,)𝐩j+nl]\displaystyle s_{c}^{*}\sum_{l=1}^{M}\left[a_{c}s_{c}\left(\hat{\mathbf{h}}_{l,*}+\tilde{\mathbf{h}}_{l,*}\right)\mathbf{p}_{c}+\sum_{j=1}^{M}a_{j}s_{j}\left(\hat{\mathbf{h}}_{l,*}+\tilde{\mathbf{h}}_{l,*}\right)\mathbf{p}_{j}+n_{l}\right]
+s1[acsc(𝐡^1,+𝐡~1,)𝐩c+l=1Malsl(𝐡^1,+𝐡~1,)𝐩l+n1]\displaystyle+s_{1}^{*}\left[a_{c}s_{c}\left(\hat{\mathbf{h}}_{1,*}+\tilde{\mathbf{h}}_{1,*}\right)\mathbf{p}_{c}+\sum_{l=1}^{M}a_{l}s_{l}\left(\hat{\mathbf{h}}_{1,*}+\tilde{\mathbf{h}}_{1,*}\right)\mathbf{p}_{l}+n_{1}\right]
++sM[acsc(𝐡^M,+𝐡~M,)𝐩c+l=1Malsl(𝐡^M,+𝐡~M,)𝐩l+nM].\displaystyle+\cdots+s_{M}^{*}\left[a_{c}s_{c}\left(\hat{\mathbf{h}}_{M,*}+\tilde{\mathbf{h}}_{M,*}\right)\mathbf{p}_{c}+\sum_{l=1}^{M}a_{l}s_{l}\left(\hat{\mathbf{h}}_{M,*}+\tilde{\mathbf{h}}_{M,*}\right)\mathbf{p}_{l}+n_{M}\right]. (54)

By evaluating the expected value of the different terms in (54) we get the following quantities:

𝔼[siyi|𝐇^]\displaystyle\mathbb{E}\left[s_{i}^{*}y_{i}|\hat{\mathbf{H}}\right] =𝔼[si{acsc(𝐡^i,+𝐡~i,)𝐩c+l=1Malsl(𝐡^i,+𝐡~i,)𝐩l+ni}|𝐇^],\displaystyle=\mathbb{E}\left[s_{i}^{*}\left\{a_{c}s_{c}\left(\hat{\mathbf{h}}_{i,*}+\tilde{\mathbf{h}}_{i,*}\right)\mathbf{p}_{c}+\sum_{l=1}^{M}a_{l}s_{l}\left(\hat{\mathbf{h}}_{i,*}+\tilde{\mathbf{h}}_{i,*}\right)\mathbf{p}_{l}+n_{i}\right\}\lvert\hat{\mathbf{H}}\right],
=aiϕ^(i,i).\displaystyle=a_{i}\hat{\phi}^{\left(i,i\right)}. (55)
𝔼[scyc|𝐇^]\displaystyle\mathbb{E}\left[s_{c}^{*}y_{c}|\hat{\mathbf{H}}\right] =𝔼[scl=1M{acsc(𝐡^l,+𝐡~l,)𝐩c+j=1Majsj(𝐡^l,+𝐡~l,)𝐩j+nl}|𝐇^],\displaystyle=\mathbb{E}\left[s_{c}^{*}\sum_{l=1}^{M}\left\{a_{c}s_{c}\left(\hat{\mathbf{h}}_{l,*}+\tilde{\mathbf{h}}_{l,*}\right)\mathbf{p}_{c}+\sum_{j=1}^{M}a_{j}s_{j}\left(\hat{\mathbf{h}}_{l,*}+\tilde{\mathbf{h}}_{l,*}\right)\mathbf{p}_{j}+n_{l}\right\}\lvert\hat{\mathbf{H}}\right],
=aci=1Mϕ^(i,c),\displaystyle=a_{c}\sum_{i=1}^{M}\hat{\phi}^{\left(i,c\right)}, (56)

where ϕ^(i,q)=𝐡^i,𝐩q\hat{\phi}^{\left(i,q\right)}=\hat{\mathbf{h}}_{i,*}\mathbf{p}_{q} and ϕ^(i,c)=𝐡^i,𝐩c\hat{\phi}^{\left(i,c\right)}=\hat{\mathbf{h}}_{i,*}\mathbf{p}_{c}. These expressions allow us to compute T2T_{2}, which is expressed by

𝔼[𝐬(RS)H𝐲|𝐇^]=aci=1Mϕ^(i,c)+j=1Majϕ^(j,j).\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{y}^{\prime}|\hat{\mathbf{H}}\right]=a_{c}\sum_{i=1}^{M}\hat{\phi}^{\left(i,c\right)}+\sum_{j=1}^{M}a_{j}\hat{\phi}^{\left(j,j\right)}. (57)

The third term can be calculated in a similar manner and is given by

𝔼[𝐲H𝐬(RS)\displaystyle\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{s}^{\left(\text{RS}\right)}\right. |𝐇^]=aci=1Mϕ^(i,c)+j=1Majϕ^(j,j).\displaystyle\left.|\hat{\mathbf{H}}\right]=a_{c}\sum_{i=1}^{M}\hat{\phi}^{\left(i,c\right)^{*}}+\sum_{j=1}^{M}a_{j}\hat{\phi}^{\left(j,j\right)^{*}}. (58)

The last term of equation (52) requires the computation of several quantities. Let us first consider the following quantity:

𝐲H𝐲=\displaystyle\mathbf{y^{\prime}}^{H}\mathbf{y}^{\prime}= ycyc+y1y1++yMyM=(l=1Myl)(j=1Myj)+i=1Myiyi.\displaystyle y_{c}^{*}y_{c}+y_{1}^{*}y_{1}+\cdots+y_{M}^{*}y_{M}=\left(\sum_{l=1}^{M}y_{l}^{*}\right)\left(\sum_{j=1}^{M}y_{j}\right)+\sum_{i=1}^{M}y_{i}^{*}y_{i}. (59)

Taking the expected value on the last equation results in

𝔼[𝐲H𝐲|𝐇^]=\displaystyle\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{y}^{\prime}|\hat{\mathbf{H}}\right]= l=1Mj=1M𝔼[ylyj|𝐇^]+i=1M𝔼[yiyi|𝐇^],\displaystyle\sum_{l=1}^{M}\sum_{j=1}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}|\hat{\mathbf{H}}\right]+\sum_{i=1}^{M}\mathbb{E}\left[y_{i}^{*}y_{i}|\hat{\mathbf{H}}\right],
=\displaystyle= l=1Mj=1jlM𝔼[ylyj|𝐇^]+2i=1M𝔼[yiyi|𝐇^],\displaystyle\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}|\hat{\mathbf{H}}\right]+2\sum_{i=1}^{M}\mathbb{E}\left[y_{i}^{*}y_{i}|\hat{\mathbf{H}}\right], (60)
𝔼[yiyi|𝐇^]=\displaystyle\mathbb{E}\left[y_{i}^{*}y_{i}\lvert\hat{\mathbf{H}}\right]= 𝔼[|acsc𝐡i,𝐩c+q=1Maqsq𝐡i,𝐩q+ni|2|𝐇^],\displaystyle\mathbb{E}\left[\left\lvert a_{c}s_{c}\mathbf{h}_{i,*}\mathbf{p}_{c}+\sum_{q=1}^{M}a_{q}s_{q}\mathbf{h}_{i,*}\mathbf{p}_{q}+n_{i}\right\rvert^{2}\rvert\hat{\mathbf{H}}\right],
=\displaystyle= 𝔼[ac2|sc|2|ϕ^(i,c)+ϕ~(i,c)|2+q=1Maq|sq|2|ϕ^(i,q)+ϕ~(i,q)|2+|ni|2|𝐇^],\displaystyle\mathbb{E}\left[a_{c}^{2}\lvert s_{c}\rvert^{2}\left\lvert\hat{\phi}^{\left(i,c\right)}+\tilde{\phi}^{\left(i,c\right)}\right\rvert^{2}+\sum_{q=1}^{M}a_{q}\lvert s_{q}\rvert^{2}\left\lvert\hat{\phi}^{\left(i,q\right)}+\tilde{\phi}^{\left(i,q\right)}\right\rvert^{2}+\lvert n_{i}\rvert^{2}\lvert\hat{\mathbf{H}}\right], (61)

where ϕ~(i,c)=𝐡~i,𝐩c\tilde{\phi}^{\left(i,c\right)}=\tilde{\mathbf{h}}_{i,*}\mathbf{p}_{c} and ϕ~(i,q)=𝐡~i,𝐩q\tilde{\phi}^{\left(i,q\right)}=\tilde{\mathbf{h}}_{i,*}\mathbf{p}_{q}. Expanding the terms of the last equation results in

𝔼[yiyi|𝐇^]=\displaystyle\mathbb{E}\left[y_{i}^{*}y_{i}\lvert\hat{\mathbf{H}}\right]= 𝔼[q=1Maq2|sq|2(|ϕ^(i,q)|2+2[ϕ^(i,q)ϕ~(i,q)]+|ϕ~(i,q)|2)|𝐇^]\displaystyle\mathbb{E}\left[\sum_{q=1}^{M}a_{q}^{2}\lvert s_{q}\rvert^{2}\left(\lvert\hat{\phi}^{\left(i,q\right)}\rvert^{2}+2\Re\left[\hat{\phi}^{\left(i,q\right)^{*}}\tilde{\phi}^{\left(i,q\right)}\right]+\lvert\tilde{\phi}^{\left(i,q\right)}\rvert^{2}\right)\lvert\hat{\mathbf{H}}\right]
+𝔼[ac2|sc|2(|ϕ^(i,c)|2+2[ϕ^(i,c)ϕ~(i,c)]+|ϕ~(i,c)|2)|𝐇^]+σn2.\displaystyle+\mathbb{E}\left[a_{c}^{2}\lvert s_{c}\rvert^{2}\left(\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+2\Re\left[\hat{\phi}^{\left(i,c\right)^{*}}\tilde{\phi}^{\left(i,c\right)}\right]+\lvert\tilde{\phi}^{\left(i,c\right)}\rvert^{2}\right)|\hat{\mathbf{H}}\right]+\sigma_{n}^{2}. (62)

Since the entries of 𝐡~i,i\tilde{\mathbf{h}}_{i,*}~{}~{}\forall i are uncorrelated with zero mean and also independent from 𝐬(RS)\mathbf{s}^{\left(\text{RS}\right)}, we have

𝔼[yiyi|𝐇^]=\displaystyle\mathbb{E}\left[y_{i}^{*}y_{i}|\hat{\mathbf{H}}\right]= ac2(|ϕ^(i,c)|2+𝔼[|ϕ~(i,c)|2|𝐇^])+q=1Maq2(|ϕ^(i,q)|2+𝔼[|ϕ~(i,q)|2|𝐇^])+σn2.\displaystyle a_{c}^{2}\left(\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+\mathbb{E}\left[\lvert\tilde{\phi}^{\left(i,c\right)}\rvert^{2}\lvert\hat{\mathbf{H}}\right]\right)+\sum_{q=1}^{M}a_{q}^{2}\left(\lvert\hat{\phi}^{\left(i,q\right)}\rvert^{2}+\mathbb{E}\left[\lvert\tilde{\phi}^{\left(i,q\right)}\rvert^{2}\lvert\hat{\mathbf{H}}\right]\right)+\sigma_{n}^{2}. (63)

Note that |ϕ~(i,c)|2\lvert\tilde{\phi}^{\left(i,c\right)}\rvert^{2} and |ϕ~(i,q)|2\lvert\tilde{\phi}^{\left(i,q\right)}\rvert^{2} are independent from 𝐇^\hat{\mathbf{H}}. Thus, we get

𝔼[|ϕ~(i,c)|2]=\displaystyle\mathbb{E}\left[\lvert\tilde{\phi}^{\left(i,c\right)}\rvert^{2}\right]= |p1(c)|2𝔼[h~i,1h~i,1]+|p2(c)|2𝔼[h~i,2h~i,2]++|pNt(c)|2𝔼[h~i,Nth~i,Nt],\displaystyle\lvert p^{\left(c\right)}_{1}\rvert^{2}\mathbb{E}\left[\tilde{h}_{i,1}^{*}\tilde{h}_{i,1}\right]+\lvert p^{\left(c\right)}_{2}\rvert^{2}\mathbb{E}\left[\tilde{h}_{i,2}^{*}\tilde{h}_{i,2}\right]+\cdots+\lvert p^{\left(c\right)}_{N_{t}}\rvert^{2}\mathbb{E}\left[\tilde{h}_{i,N_{t}}^{*}\tilde{h}_{i,N_{t}}\right],
=\displaystyle= |p1(c)|2σe,i2+|p2(c)|2σe,i2+|pNt(c)|2σe,i2,\displaystyle\lvert p^{\left(c\right)}_{1}\rvert^{2}\sigma_{e,i}^{2}+\lvert p^{\left(c\right)}_{2}\rvert^{2}\sigma_{e,i}^{2}+\cdots\lvert p^{\left(c\right)}_{N_{t}}\rvert^{2}\sigma_{e,i}^{2},
=\displaystyle= σe,i2𝐩c2,\displaystyle\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}, (64)

and similarly 𝔼[|ϕ~(i,q)|2]=σe,i2𝐩q2.\mathbb{E}\left[\lvert\tilde{\phi}^{\left(i,q\right)}\rvert^{2}\right]=\sigma_{e,i}^{2}\lVert\mathbf{p}_{q}\rVert^{2}. Then, (63) turns into

𝔼[yiyi|𝐇^]=\displaystyle\mathbb{E}\left[y_{i}^{*}y_{i}|\hat{\mathbf{H}}\right]= ac2(|ϕ^(i,c)|2+σe,i2𝐩c2)+j=1Kaj2(|ϕ^(i,j)|2+σe,i2𝐩j2)+σn2.\displaystyle a_{c}^{2}\left(\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right)+\sum_{j=1}^{K}a_{j}^{2}\left(\lvert\hat{\phi}^{\left(i,j\right)}\rvert^{2}+\sigma_{e,i}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right)+\sigma_{n}^{2}. (65)

Let us now evaluate the expected value of ylyjy_{l}^{*}y_{j} when ljl\neq j, which results in

𝔼[ylyj|𝐇^]=\displaystyle\mathbb{E}\left[y_{l}^{*}y_{j}\lvert\hat{\mathbf{H}}\right]= 𝔼[(acsc𝐡l,𝐩c+q=1Maqsq𝐡l,𝐩q+nl)×(acsc𝐡j,𝐩c+r=1Marsr𝐡j,𝐩r+nj)|𝐇^],\displaystyle\mathbb{E}\left[\left(a_{c}s_{c}\mathbf{h}_{l,*}\mathbf{p}_{c}+\sum_{q=1}^{M}a_{q}s_{q}\mathbf{h}_{l,*}\mathbf{p}_{q}+n_{l}\right)^{*}\right.\times\left.\left(a_{c}s_{c}\mathbf{h}_{j,*}\mathbf{p}_{c}+\sum_{r=1}^{M}a_{r}s_{r}\mathbf{h}_{j,*}\mathbf{p}_{r}+n_{j}\right)\lvert\hat{\mathbf{H}}\right],
=\displaystyle= q=1Maq2𝔼[ϕ^(l,q)ϕ^(j,q)+ϕ^(l,q)ϕ~(j,q)+ϕ~(l,q)ϕ^(j,q)+ϕ~(l,q)ϕ~(j,q)|𝐇^].\displaystyle\sum_{q=1}^{M}a_{q}^{2}\mathbb{E}\left[\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}+\hat{\phi}^{\left(l,q\right)^{*}}\tilde{\phi}^{\left(j,q\right)}+\tilde{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}+\tilde{\phi}^{\left(l,q\right)^{*}}\tilde{\phi}^{\left(j,q\right)}\lvert\hat{\mathbf{H}}\right].
+ac2𝔼[ϕ^(l,c)ϕ^(j,c)+ϕ^(l,c)ϕ~(j,c)+ϕ~(l,c)ϕ^(j,c)+ϕ~(l,c)ϕ~(j,c)|𝐇^].\displaystyle+a_{c}^{2}\mathbb{E}\left[\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}+\hat{\phi}^{\left(l,c\right)^{*}}\tilde{\phi}^{\left(j,c\right)}+\tilde{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}+\tilde{\phi}^{\left(l,c\right)^{*}}\tilde{\phi}^{\left(j,c\right)}\lvert\hat{\mathbf{H}}\right]. (66)

Remark that 𝐡~l\tilde{\mathbf{h}}_{l} and 𝐡~j\tilde{\mathbf{h}}_{j} are independent lj\forall~{}~{}l\neq j with zero mean. Thus, the last equation is reduced to

𝔼[ylyj|𝐇^]=\displaystyle\mathbb{E}\left[y_{l}^{*}y_{j}|\hat{\mathbf{H}}\right]= ac2ϕ^(l,c)ϕ^(j,c)+q=1Maq2ϕ^(l,q)ϕ^(j,q).\displaystyle a_{c}^{2}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}+\sum_{q=1}^{M}a_{q}^{2}\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}. (67)

Equations (65) allow us to compute the second term of equation (60), which is given by

i=1M𝔼[yiyi|𝐇^]=\displaystyle\sum_{i=1}^{M}\mathbb{E}\left[y_{i}^{*}y_{i}|\right.\left.\hat{\mathbf{H}}\right]= j=1Maj2(l=1M|ϕ^(l,j)|2+Mσei2𝐩j2)+ac2(i=1M|ϕ^(i,c)|2+Mσe,i2𝐩c2)+Mσn2.\displaystyle\sum_{j=1}^{M}a_{j}^{2}\left(\sum_{l=1}^{M}\lvert\hat{\phi}^{\left(l,j\right)}\rvert^{2}+M\sigma_{e_{i}}^{2}\lVert\mathbf{p}_{j}\rVert^{2}\right)+a_{c}^{2}\left(\sum_{i=1}^{M}\lvert\hat{\phi}^{\left(i,c\right)}\rvert^{2}+M\sigma_{e,i}^{2}\lVert\mathbf{p}_{c}\rVert^{2}\right)+M\sigma_{n}^{2}. (68)

On the other hand, (67) allow us to obtain the first term of (60), which results in

l=1Mj=1jlM𝔼[ylyj|𝐇^]=l=1Mj=1jlM(ac2ϕ^(l,c)ϕ^(j,c)+q=1Maq2ϕ^(l,q)ϕ^(j,q))\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}|\hat{\mathbf{H}}\right]=\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\left(a_{c}^{2}\hat{\phi}^{\left(l,c\right)^{*}}\hat{\phi}^{\left(j,c\right)}+\sum_{q=1}^{M}a_{q}^{2}\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}\right) (69)

Applying the property ϕ^(l,q)ϕ^(j,q)+ϕ^(l,q)ϕ^(j,q)=2[ϕ^(l,q)ϕ^(j,q)]\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}+\hat{\phi}^{\left(l,q\right)}\hat{\phi}^{\left(j,q\right)^{*}}=2\Re\left[\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}\right], we can simplify half of the sums from the triple summation, i.e.,

l=1Mj=1jlMq=1Maq2ϕ^(l,q)ϕ^(j,q)=2l=1M1j=i+1Mq=1Maq2[ϕ^(l,q)ϕ^(j,q)].\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\sum_{q=1}^{M}a_{q}^{2}\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}=2\sum_{l=1}^{M-1}\sum_{j=i+1}^{M}\sum_{q=1}^{M}a_{q}^{2}\Re\left[\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}\right]. (70)

It follows that

l=1Mj=1jlM𝔼[ylyj|𝐇^]=\displaystyle\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}|\hat{\mathbf{H}}\right]= 2l=1M1j=i+1Mq=1Mar2[ϕ^(l,q)ϕ^(j,q)]+l=1Mj=1jlMac2ϕ^(l,c)ϕ^(j,c).\displaystyle 2\sum_{l=1}^{M-1}\sum_{j=i+1}^{M}\sum_{q=1}^{M}a_{r}^{2}\Re\left[\hat{\phi}^{\left(l,q\right)^{*}}\hat{\phi}^{\left(j,q\right)}\right]+\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}^{2}\hat{\phi}^{\left(l,c\right)}\hat{\phi}^{\left(j,c\right)}. (71)

By using (68) and (71) in (60) and then substituting (53) (57), (58), and (60) in (52) we can calculate the MSE, which is given by (9). This concludes the proof.

Appendix B Proof of the MSE for the APA

Here, we describe in detail how to obtain the MSE employed in the APA algorithm. Let us first consider the MSE, which is given by

𝔼[ε]=\displaystyle\mathbb{E}\left[\varepsilon\right]= 𝔼[𝐬(RS)H𝐬(RS)]𝔼[𝐬(RS)H𝐲]𝔼[𝐲H𝐬(RS)]+𝔼[𝐲H𝐲].\displaystyle{\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{s}^{\left(\text{RS}\right)}\right]}-{\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{y}^{\prime}\right]}-{\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{s}^{\left(\text{RS}\right)}\right]}+{\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{y}^{\prime}\right]}. (72)

The first term of (72) is computed identically to (53).

By taking the expected value of the second term in (72) and expanding the equation, we have

𝔼[𝐬(RS)H𝐲]=\displaystyle\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{y}^{\prime}\right]= ac𝔼[scsc]l=1M𝐡l,𝐩c+l=1M𝔼[scnl]+l=1M𝐡l,j=1Maj𝔼[scsj]𝐩j\displaystyle a_{c}\mathbb{E}\left[s_{c}^{*}s_{c}\right]\sum_{l=1}^{M}\mathbf{h}_{l,*}\mathbf{p}_{c}+\sum_{l=1}^{M}\mathbb{E}\left[s_{c}^{*}n_{l}\right]+\sum_{l=1}^{M}\mathbf{h}_{l,*}\sum_{j=1}^{M}a_{j}\mathbb{E}\left[s_{c}^{*}s_{j}\right]\mathbf{p}_{j}
+acl=1M𝔼[slsc]𝐡l,𝐩c+q=1Ml=1Mal𝔼[sqsl]𝐡q,𝐩l+l=1M𝔼[slnl].\displaystyle+a_{c}\sum_{l=1}^{M}\mathbb{E}\left[s_{l}^{*}s_{c}\right]\mathbf{h}_{l,*}\mathbf{p}_{c}+\sum_{q=1}^{M}\sum_{l=1}^{M}a_{l}\mathbb{E}\left[s_{q}^{*}s_{l}\right]\mathbf{h}_{q,*}\mathbf{p}_{l}+\sum_{l=1}^{M}\mathbb{E}\left[s_{l}^{*}n_{l}\right]. (73)

Since the symbols are uncorrelated, equation (73) is reduced to

𝔼[𝐬(RS)H𝐲]=acj=1M𝐡j,𝐩c+l=1Mal𝐡l,𝐩l.\displaystyle\mathbb{E}\left[\mathbf{s}^{\left(\text{RS}\right)^{H}}\mathbf{y}^{\prime}\right]=a_{c}\sum_{j=1}^{M}\mathbf{h}_{j,*}\mathbf{p}_{c}+\sum_{l=1}^{M}a_{l}\mathbf{h}_{l,*}\mathbf{p}_{l}. (74)

The third term of (72) can be computed in a similar way as the second term, which lead us to

𝔼[𝐲H𝐬(RS)]=acj=1M(𝐡l,𝐩c)+l=1Mal(𝐡l,𝐩l).\displaystyle\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{s}^{\left(\text{RS}\right)}\right]=a_{c}\sum_{j=1}^{M}\left(\mathbf{h}_{l,*}\mathbf{p}_{c}\right)^{*}+\sum_{l=1}^{M}a_{l}\left(\mathbf{h}_{l,*}\mathbf{p}_{l}\right)^{*}. (75)

The last term of (72) is equal to

𝔼[𝐲H𝐲]=l=1Mj=1jlM𝔼[ylyj]+2i=1M𝔼[yiyi],\displaystyle\mathbb{E}\left[\mathbf{y^{\prime}}^{H}\mathbf{y}^{\prime}\right]=\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}\right]+2\sum_{i=1}^{M}\mathbb{E}\left[y_{i}^{*}y_{i}\right], (76)

Let us first compute the quantity given by

𝔼[yiyi]=ac2|𝐡i,𝐩c|2+l=1Mal2|𝐡i,𝐩l|2+σn2.\mathbb{E}\left[y_{i}^{*}y_{i}\right]=a_{c}^{2}\lvert\mathbf{h}_{i,*}\mathbf{p}_{c}\rvert^{2}+\sum_{l=1}^{M}a_{l}^{2}\lvert\mathbf{h}_{i,*}\mathbf{p}_{l}\rvert^{2}+\sigma_{n}^{2}. (77)

Additionally, we know that

𝔼[yiyj]=\displaystyle\mathbb{E}\left[y_{i}^{*}y_{j}\right]= 𝔼[(acsc𝐡i,𝐩c+q=1Maqsq𝐡i,𝐩q+ni)×(acsc𝐡j,𝐩c+l=1Malsl𝐡j,𝐩l+nj)],\displaystyle\mathbb{E}\left[\left(a_{c}s_{c}\mathbf{h}_{i,*}\mathbf{p}_{c}+\sum_{q=1}^{M}a_{q}s_{q}\mathbf{h}_{i,*}\mathbf{p}_{q}+n_{i}\right)^{*}\right.\times\left.\left(a_{c}s_{c}\mathbf{h}_{j,*}\mathbf{p}_{c}+\sum_{l=1}^{M}a_{l}s_{l}\mathbf{h}_{j,*}\mathbf{p}_{l}+n_{j}\right)\right],
=\displaystyle= ac2(𝐡i,𝐩c)(𝐡j,𝐩c)+l=1Mal2(𝐡i,𝐩l)(𝐡j,𝐩l),\displaystyle a_{c}^{2}\left(\mathbf{h}_{i,*}\mathbf{p}_{c}\right)^{*}\left(\mathbf{h}_{j,*}\mathbf{p}_{c}\right)+\sum_{l=1}^{M}a_{l}^{2}\left(\mathbf{h}_{i,*}\mathbf{p}_{l}\right)^{*}\left(\mathbf{h}_{j,*}\mathbf{p}_{l}\right), (78)

for all iji\neq j. From the last equation, we have

i=1Mj=1jlM𝔼[ylyj]=\displaystyle\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}\right]= i=1Mj=1jlM[ac2(𝐡i,𝐩c)(𝐡j,𝐩c)+l=1Mal2(𝐡i,𝐩l)(𝐡j,𝐩l)],\displaystyle\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\left[a_{c}^{2}\left(\mathbf{h}_{i,*}\mathbf{p}_{c}\right)^{*}\left(\mathbf{h}_{j,*}\mathbf{p}_{c}\right)+\sum_{l=1}^{M}a_{l}^{2}\left(\mathbf{h}_{i,*}\mathbf{p}_{l}\right)^{*}\left(\mathbf{h}_{j,*}\mathbf{p}_{l}\right)\right],
=\displaystyle= i=1Mj=1jlMac2(𝐡i,𝐩c)(𝐡j,𝐩c)+i=1Mj=1jlMl=1Mal2(𝐡i,𝐩l)(𝐡j,𝐩l),\displaystyle\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}^{2}\left(\mathbf{h}_{i,*}\mathbf{p}_{c}\right)^{*}\left(\mathbf{h}_{j,*}\mathbf{p}_{c}\right)+\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\sum_{l=1}^{M}a_{l}^{2}\left(\mathbf{h}_{i,*}\mathbf{p}_{l}\right)^{*}\left(\mathbf{h}_{j,*}\mathbf{p}_{l}\right),
=\displaystyle= i=1Mj=1jlMac2ϕ(i,c)ϕ(j,c)+i=1Mj=1jlMl=1Mal2ϕ(i,l)ϕ(j,l),\displaystyle\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}^{2}\phi^{\left(i,c\right)^{*}}\phi^{\left(j,c\right)}+\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\sum_{l=1}^{M}a_{l}^{2}\phi^{\left(i,l\right)^{*}}\phi^{\left(j,l\right)}, (79)

where we define ϕ(i,c)=𝐡l,𝐩c\phi^{\left(i,c\right)}=\mathbf{h}_{l,*}\mathbf{p}_{c} and ϕ(i,l)=𝐡i,𝐩l\phi^{\left(i,l\right)}=\mathbf{h}_{i,*}\mathbf{p}_{l} for all i,l[1,M]i,l\in\left[1,M\right]. Applying the property ϕ(i,l)ϕ(j,l)+ϕ(i,l)ϕ(j,l)=2[ϕ(i,l)ϕ(j,l)]\phi^{\left(i,l\right)^{*}}\phi^{\left(j,l\right)}+\phi^{\left(i,l\right)}\phi^{\left(j,l\right)^{*}}=2\Re\left[\phi^{\left(i,l\right)^{*}}\phi^{\left(j,l\right)}\right], we can simplify half of the sums from the triple summation, i.e.,

i=1Mj=1jlMl=1Mal2ϕ(i,l)ϕ(j,l)=2i=1M1q=i+1Mr=1Mar2[ϕ(i,r)ϕ(q,r)].\sum_{i=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\sum_{l=1}^{M}a_{l}^{2}\phi^{\left(i,l\right)^{*}}\phi^{\left(j,l\right)}=2\sum_{i=1}^{M-1}\sum_{q=i+1}^{M}\sum_{r=1}^{M}a_{r}^{2}\Re\left[\phi^{\left(i,r\right)^{*}}\phi^{\left(q,r\right)}\right]. (80)

The final step to obtain the last term of (72) is to employ (77), (79), and (80) to compute the following quantities:

i=1M𝔼[yiyi]=l=1Mac2|𝐡l,𝐩c|2+i=1Mj=1Maj2|𝐡i,𝐩j|2+Mσn2,\sum_{i=1}^{M}\mathbb{E}\left[y_{i}^{*}y_{i}\right]=\sum_{l=1}^{M}a_{c}^{2}\lvert\mathbf{h}_{l,*}\mathbf{p}_{c}\rvert^{2}+\sum_{i=1}^{M}\sum_{j=1}^{M}a_{j}^{2}\lvert\mathbf{h}_{i,*}\mathbf{p}_{j}\rvert^{2}+M\sigma_{n}^{2}, (81)
l=1Mj=1jlM𝔼[ylyj]=\displaystyle\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}\mathbb{E}\left[y_{l}^{*}y_{j}\right]= 2i=1M1q=i+1Mr=1Mar2[ϕ(i,r)ϕ(q,r)]+l=1Mj=1jlMac2ϕ(l,c)ϕ(j,c).\displaystyle 2\sum_{i=1}^{M-1}\sum_{q=i+1}^{M}\sum_{r=1}^{M}a_{r}^{2}\Re\left[\phi^{\left(i,r\right)^{*}}\phi^{\left(q,r\right)}\right]+\sum_{l=1}^{M}\sum\limits_{\begin{subarray}{c}j=1\\ j\neq l\end{subarray}}^{M}a_{c}^{2}\phi^{\left(l,c\right)^{*}}\phi^{\left(j,c\right)}. (82)

By using (81) and (82) in (76) and then substituting (53), (74), (75), (76) in (72) we get the MSE in (14).

References

  • [1] R. C. de Lamare, “Massive mimo systems: Signal processing challenges and future trends,” URSI Radio Science Bulletin, vol. 2013, no. 347, pp. 8–20, 2013.
  • [2] M. Joham, W. Utschick, and J. A. Nossek, “Linear transmit processing in MIMO communications systems,” IEEE Transactions on Signal Processing, vol. 53, no. 8, pp. 2700–2712, Aug. 2005.
  • [3] W. Zhang, H. Ren, C. Pan, M. Chen, R. C. de Lamare, B. Du, and J. Dai, “Large-scale antenna systems with ul/dl hardware mismatch: Achievable rates analysis and calibration,” IEEE Transactions on Communications, vol. 63, no. 4, pp. 1216–1229, 2015.
  • [4] K. Zu, R. C. de Lamare, and M. Haardt, “Generalized design of low-complexity block diagonalization type precoding algorithms for multiuser mimo systems,” IEEE Transactions on Communications, vol. 61, no. 10, pp. 4232–4242, 2013.
  • [5] W. Zhang, R. C. de Lamare, C. Pan, M. Chen, J. Dai, B. Wu, and X. Bao, “Widely linear precoding for large-scale mimo with iqi: Algorithms and performance analysis,” IEEE Transactions on Wireless Communications, vol. 16, no. 5, pp. 3298–3312, 2017.
  • [6] Y. Cai, R. C. de Lamare, L.-L. Yang, and M. Zhao, “Robust mmse precoding based on switched relaying and side information for multiuser mimo relay systems,” IEEE Transactions on Vehicular Technology, vol. 64, no. 12, pp. 5677–5687, 2015.
  • [7] K. Zu, R. C. de Lamare, and M. Haardt, “Multi-branch Tomlinson-Harashima precoding design for MU-MIMO systems: Theory and algorithms,” IEEE Transactions on Communications, vol. 62, no. 3, pp. 939–951, Mar. 2014.
  • [8] C. Peel, B. Hochwald, and A. Swindlehurst, “A vector-perturbation technique for near-capacity multiantenna multiuser communication— part II: Perturbation,” IEEE Transactions on Communications, vol. 53, no. 1, pp. 537–544, Mar. 2005.
  • [9] L. T. N. Landau and R. C. de Lamare, “Branch-and-bound precoding for multiuser mimo systems with 1-bit quantization,” IEEE Wireless Communications Letters, vol. 6, no. 6, pp. 770–773, 2017.
  • [10] D. Tse and P. Viswanath, Fundamentals of Wireless Communications.   Cambridge University Press, 2005.
  • [11] B. Clerckx, H. Joudeh, C. Hao, M. Dai, and B. Rassouli, “Rate splitting for MIMO wireless networks: a promising PHY-layer strategy for LTE evolution,” IEEE Communications Magazine, vol. 54, no. 5, pp. 98–105, May 2016.
  • [12] T. S. Han and K. Kobayashi, “A new achievable rate region for the interference channel,” IEEE Transactions on Information Theory, vol. 27, no. 1, pp. 49–60, Jan. 1981.
  • [13] A. Carleial, “Interference channels,” IEEE Transactions on Information Theory, vol. 24, no. 1, pp. 60–70, Jan. 1978.
  • [14] A. Haghi and A. K. Khandani, “Rate splitting and successive decoding for gaussian interference channels,” IEEE Transactions on Information Theory, vol. 67, no. 3, pp. 1699–1731, 2021.
  • [15] Y. Mao, “Rate-splitting multiple access for downlink communications systems,” Ph.D. dissertation, The University of Hong Kong, 2018.
  • [16] Y. Mao and B. Clerckx, “Beyond dirty paper coding for multi-antenna broadcast channel with partial CSIT: A rate-splitting approach,” IEEE Transactions on Communications, vol. 68, no. 11, pp. 6775–6791, Nov. 2020.
  • [17] B. Clerckx, Y. Mao, R. Schober, and H. V. Poor, “Rate-splitting unifying SDMA, OMA, NOMA, and multicasting in MISO broadcast channel: A simple two-user rate analysis,” IEEE Wireless Communications Letters, vol. 9, no. 3, pp. 349–353, Mar. 2020.
  • [18] S. Naser, P. C. Sofotasios, L. Bariah, W. Jaafar, S. Muhaidat, M. Al-Qutayri, and O. A. Dobre, “Rate-splitting multiple access: Unifying NOMA and SDMA in MISO VLC channels,” IEEE Open Journal of Vehicular Technology, vol. 1, pp. 393–413, Oct. 2020.
  • [19] W. Jaafar, S. Naser, S. Muhaidat, P. C. Sofotasios, and H. Yanikomeroglu, “Multiple access in aerial networks: From orthogonal and non-orthogonal to rate-splitting,” IEEE Open Journal of Vehicular Technology, vol. 1, pp. 372–392, 2020.
  • [20] Y. Mao, O. Dizdar, B. Clerckx, R. Schober, P. Popovski, and H. V. Poor, “Rate-splitting multiple access: Fundamentals, survey, and future research trends,” arXiv:2201.03192v1 [cs.IT], Jan. 2022.
  • [21] R. C. De Lamare and R. Sampaio-Neto, “Minimum mean-squared error iterative successive parallel arbitrated decision feedback detectors for ds-cdma systems,” IEEE Transactions on Communications, vol. 56, no. 5, pp. 778–789, 2008.
  • [22] P. Li, R. C. de Lamare, and R. Fa, “Multiple feedback successive interference cancellation detection for multiuser mimo systems,” IEEE Transactions on Wireless Communications, vol. 10, no. 8, pp. 2434–2439, 2011.
  • [23] R. C. de Lamare, “Adaptive and iterative multi-branch mmse decision feedback detection algorithms for multi-antenna systems,” IEEE Transactions on Wireless Communications, vol. 12, no. 10, pp. 5294–5308, 2013.
  • [24] A. G. D. Uchoa, C. T. Healy, and R. C. de Lamare, “Iterative detection and decoding algorithms for mimo systems in block-fading channels using ldpc codes,” IEEE Transactions on Vehicular Technology, vol. 65, no. 4, pp. 2735–2741, 2016.
  • [25] Z. Shao, R. C. de Lamare, and L. T. N. Landau, “Iterative detection and decoding for large-scale multiple-antenna systems with 1-bit adcs,” IEEE Wireless Communications Letters, vol. 7, no. 3, pp. 476–479, 2018.
  • [26] S. Yang, M. Kobayashi, D. Gesbert, and X. Yi, “Degrees of freedom of time correlated MISO broadcast channel with delayed CSIT,” IEEE Transactions on Information Theory, vol. 59, no. 1, pp. 315–328, Jan. 2013.
  • [27] C. Hao, B. Rassouli, and B. Clerckx, “Achievable dof regions of mimo networks with imperfect csit,” IEEE Transactions on Information Theory, vol. 63, no. 10, pp. 6587–6606, Oct 2017.
  • [28] E. Piovano and B. Clerckx, “Optimal DoF region of the KK -user MISO BC with partial CSIT,” IEEE Communications Letters, vol. 21, no. 11, pp. 2368–2371, Nov. 2017.
  • [29] H. Joudeh, “A rate-spliting approach to multiple-antenna broadcasting,” Ph.D. dissertation, Imperial College London, 2016.
  • [30] C. Hao, Y. Wu, and B. Clerckx, “Rate analysis of two-receiver MISO broadcast channel with finite rate feedback: A rate-splitting approach,” IEEE Transactions on Communications, vol. 63, no. 9, pp. 3232–3246, Sep. 2015.
  • [31] H. Joudeh and B. Clerckx, “Rate-splitting for max-min fair multigroup multicast beamforming in overloaded systems,” IEEE Transactions on Wireless Communications, vol. 16, no. 11, pp. 7276–7289, Nov. 2017.
  • [32] C. Hao and B. Clerckx, “MISO networks with imperfect CSIT: A topological rate-splitting approach,” IEEE Transactions on CommunicationsIEEE Transactions on Communications, vol. 65, no. 5, pp. 2164–2179, May 2017.
  • [33] G. Lu, L. Li, H. Tian, and F. Qian, “MMSE-based precoding for rate splitting systems with finite feedback,” IEEE Communications Letters, vol. 22, no. 3, pp. 642–645, Mar. 2018.
  • [34] A. R. Flores, R. C. De Lamare, and B. Clerckx, “Linear precoding and stream combining for rate splitting in multiuser MIMO systems,” IEEE Communications Letters, vol. 24, no. 4, pp. 890–894, Jan. 2020.
  • [35] A. R. Flores, R. C. De Lamare, and B. Clerckx, “Tomlinson-harashima precoded rate-splitting with stream combiners for MU-MIMO systems,” IEEE Transactions on Communications, pp. 1–1, 2021.
  • [36] A. Krishnamoorthy and R. Schober, “Successive null-space precoder design for downlink MU-MIMO with rate splitting and single-stage SIC,” in 2021 IEEE International Conference on Communications Workshops (ICC Workshops), 2021, pp. 1–6.
  • [37] A. Mishra, Y. Mao, O. Dizdar, and B. Clerckx, “Rate-splitting multiple access for downlink multiuser MIMO: Precoder optimization and PHY-layer design,” IEEE Transactions on Communications, vol. 70, no. 2, pp. 874–890, Feb. 2022.
  • [38] Z. Li, C. Ye, Y. Cui, S. Yang, and S. Shamai, “Rate splitting for multi-antenna downlink: Precoder design and practical implementation,” IEEE Journal on Selected Areas in Communications, vol. 38, no. 8, pp. 1910–1924, 2020.
  • [39] Z. Luo and S. Zhang, “Dynamic spectrum management: Complexity and duality,” IEEE Journal of Selected Topics in Signal Processing, vol. 2, no. 1, pp. 57–73, Feb. 2008.
  • [40] Y. Liu, Y. Dai, and Z. Luo, “Coordinated beamforming for MISO interference channel: Complexity analysis and efficient algorithms,” IEEE Transactions on Signal Processing, vol. 59, no. 3, pp. 1142–1157, Mar. 2011.
  • [41] Y. Mao, B. Clerckx, and V. Li, “Rate-splitting multiple access for downlink communication systems: Bridging, generalizing and outperforming SDMA and NOMA,” EURASIP Journal on Wireless Communications and Networking, vol. 2018, no. 1, p. 133, May 2018.
  • [42] H. Joudeh and B. Clerckx, “Sum-rate maximization for linearly precoded downlink multiuser MISO systems with partial CSIT: A rate-splitting approach,” IEEE Transactions on Communications, vol. 64, no. 11, pp. 4847–4861, Nov. 2016.
  • [43] C. Kaulich, M. Joham, and W. Utschick, “Rate-splitting for the weighted sum rate maximization under minimum rate constraints in the MIMO BC,” in 2021 IEEE International Conference on Communications Workshops (ICC Workshops), 2021, pp. 1–6.
  • [44] S. S. Christensen, R. Agarwal, E. De Carvalho, and J. M. Cioffi, “Weighted sum-rate maximization using weighted MMSE for MIMO-BC beamforming design,” IEEE Transactions on Wireless Communications, vol. 7, no. 12, pp. 4792–4799, Dec. 2008.
  • [45] M. Dai, B. Clerckx, D. Gesbert, and G. Caire, “A rate splitting strategy for massive mimo with imperfect csit,” IEEE Transactions on Wireless Communications, vol. 15, no. 7, pp. 4611–4624, Jul. 2016.
  • [46] A. R. Flores and R. C. De Lamare, “Robust and adaptive power allocation techniques for rate splitting based mu-mimo systems,” IEEE Transactions on Communications, pp. 1–1, 2022.
  • [47] D. P. Bertsekas, N. Angelia, and A. E. Ozdaglar, Convex Analysis and. Optimization.   Athena Scientific, Belmont, MA:, 2003.
  • [48] R. C. de Lamare and R. Sampaio-Neto, “Adaptive reduced-rank processing based on joint and iterative interpolation, decimation, and filtering,” IEEE Transactions on Signal Processing, vol. 57, no. 7, pp. 2503–2514, 2009.
  • [49] R. C. de Lamare and P. S. R. Diniz, “Set-membership adaptive algorithms based on time-varying error bounds for cdma interference suppression,” IEEE Transactions on Vehicular Technology, vol. 58, no. 2, pp. 644–654, 2009.
  • [50] T. Wang, R. C. de Lamare, and P. D. Mitchell, “Low-complexity set-membership channel estimation for cooperative wireless sensor networks,” IEEE Transactions on Vehicular Technology, vol. 60, no. 6, pp. 2594–2607, 2011.
  • [51] H. Ruan and R. C. de Lamare, “Robust adaptive beamforming using a low-complexity shrinkage-based mismatch estimation algorithm,” IEEE Signal Processing Letters, vol. 21, no. 1, pp. 60–64, 2014.
  • [52] ——, “Robust adaptive beamforming based on low-rank and cross-correlation techniques,” IEEE Transactions on Signal Processing, vol. 64, no. 15, pp. 3919–3932, 2016.
  • [53] N. Yousef and A. Sayed, “A unified approach to the steady-state and tracking analyses of adaptive filters,” IEEE Transactions on Signal Processing, vol. 49, no. 2, pp. 314–324, 2001.
  • [54] A. Sayed, Fundamentals of Adaptive Filtering.   Fundamentals of Adaptive Filtering, 2003.