This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Optimal Beam Association for High Mobility mmWave Vehicular Networks: Lightweight Parallel Reinforcement Learning Approach

Nguyen Van Huynh, Diep N. Nguyen, Dinh Thai Hoang, and Eryk Dutkiewicz The authors are with the School of Electrical and Data Engineering, University of Technology Sydney, Sydney, NSW 2007, Australia. E-mails: huynh.nguyenvan@student.uts.edu.au, {Diep.Nguyen, Hoang.Dinh, and Eryk.Dutkiewicz}@uts.edu.au.Preliminary results in this paper were presented at the IEEE GLOBECOM Conference, 2020 [1].
Abstract

In intelligent transportation systems (ITS), vehicles are expected to feature with advanced applications and services which demand ultra-high data rates and low-latency communications. For that, the millimeter wave (mmWave) communication has been emerging as a very promising solution. However, incorporating the mmWave into ITS is particularly challenging due to the high mobility of vehicles and the inherent sensitivity of mmWave beams to dynamic blockages. This article addresses these problems by developing an optimal beam association framework for mmWave vehicular networks under high mobility. Specifically, we use the semi-Markov decision process to capture the dynamics and uncertainty of the environment. The Q-learning algorithm is then often used to find the optimal policy. However, Q-learning is notorious for its slow-convergence. Instead of adopting deep reinforcement learning structures (like most works in the literature), we leverage the fact that there are usually multiple vehicles on the road to speed up the learning process. To that end, we develop a lightweight yet very effective parallel Q-learning algorithm to quickly obtain the optimal policy by simultaneously learning from various vehicles. Extensive simulations demonstrate that our proposed solution can increase the data rate by 47% and reduce the disconnection probability by 29% compared to other solutions.

Index Terms:
millimeter wave, vehicular networks, high mobility, handover, beam selection, parallel Q-learning.

I Introduction

Over the past few years, the explosive growth of interest in intelligent transportation and vehicular communications offers a great potential to enhance traffic efficiency, improve road safety, and enable open disruptive entertainment services and autonomous driving [2][3]. These applications often require low-latency, high reliability, and especially multi-Gbps network access. For instance, Google’s self-driving car in a second can generate up to 750750 MB of data [4]. It is expected that a vehicle may produce 11 Terabyte of data in a single trip [2]. To address this critical problem, the emerging millimeter wave (mmWave) communication has been recently considered as a very promising solution [5]. Comparing with existing wireless networks, the mmWave technology operates at much higher carrier frequencies, i.e., from 30 GHz to 300 GHz. Thus, it possesses much more abundant spectrum resources, resulting in potentially extremely high data rates and low-latency communications. Nevertheless, in mmWave communications, the temporal degradation of channel quality occurs much more frequently than conventional (lower frequency) communications due to high propagation attenuation, selective directivity, and severe susceptibility to blockages, especially in vehicular communications under high mobility. This work aims to address these problems to enable mmWave communications under high mobility.

I-A Related Work and Motivation

Various solutions in the literature have been proposed to address the above inherent limitations of mmWave communications, e.g., [6]-[11]. In [6], the authors aimed to jointly optimize the transmit power, interference coordination, and beamforming to maximize the signal to interference plus noise ratio and the sum-rate. In [7], the authors aimed to enhance the users’ quality-of-experience by scheduling communication tasks on both microwave and mmWave bands. Similarly, a new protocol that enables simultaneous connections to conventional 4G cells and 5G mmWave cells is introduced in [8]. Differently, a new adaptive beamforming strategy is proposed in [10] to improve the system communication rate by taking the noise into account and selecting the beamforming vectors based on the posterior of the angle-of-arrival.

The aforementioned solutions and most others in the literature did not account for the high mobility of vehicles and dynamics of the environment. In [12], the authors proposed an adaptive channel estimation mechanism for beamforming with the aid of location information. The position information is then demonstrated as an important factor to greatly improve the initial association of vehicles to the infrastructure. Similarly, in [13], the authors introduced a beam switching mechanism in mmWave vehicle-to-infrastructure communications. In [14], a deep learning approach is proposed to predict mmWave beams and blockages with sub-6 GHz channels. Specifically, the authors studied that there are deterministic mapping from uplink sub-6 GHz channels to downlink mmWave channels. Hence, with the uplink channel knowledge, it is possible to predict/infer mmWave beams and blockages with low training overhead because sub-6 GHz channels can be estimated by using a small number of pilots. Nevertheless, it is very challenging to derive and analyze these mapping functions. To tackle this, the authors developed a deep learning model with a sufficiently large deep neural network to efficiently predict the optimal beams and blockages status.

To deal with the inherent dynamic and uncertainty of mmWave links’ quality, the authors in [15] proposed an online learning algorithm to obtain the optimal beam selection policy based on the prior environment information. This problem is first modeled as a contextual multi-armed bandit problem. Then, the learning algorithm is developed to guide the mmWave base station to select an optimal subset of beams for vehicles. In [16], the authors aimed to maximize the number of bits delivered from a base station to a mobile user by considering the trade-off between directive data transmission (DT) and directional beam training (BT) in mmWave vehicular networks. Specifically, the system is formulated as a partially observable Markov decision process with the system state defined as the position of the mobile user within the road link. Note that in both [15] and [16], only one mmWave base station is considered. In practice where multiple base stations are often in place, they can cooperate in the beam association and handoff process to achieve a globally optimal beam selection solution. An online learning approach is proposed in [17] to jointly optimize beam training, data transmission, and handover processes. Nevertheless, the effect of high mobility on the system performance has not been studied. Moreover, similar to [16], only one mobile user is considered in this paper. In [18], the authors considered the vehicle-cell association problem for mmWave vehicular networks to maximize the average rate of vehicles. Specifically, the authors first formulated the association problem as a discrete non-convex optimization problem. Then, a learning algorithm is developed to estimate the solution for the non-convex optimization problem. It is worth noting that the above works and others in the literature have not accounted for the effects of the high mobility and blockage on beam association/handover. In addition, these learning approaches do not leverage the fact that there are multiple vehicles running on the road at the same time to improve the learning efficiency.

Given the above, this work aims to develop an optimal beam association framework for mmWave vehicular communications under the high mobility of vehicles and the uncertainty of blockages. In particular, to capture the dynamics of blockages, channel quality, and mobility, we first model the problem as a semi-Markov decision process (SMDP). The Q-learning (QL) algorithm is then often adopted to solve the optimization problem the underlying SMDP. Nevertheless, the QL algorithm is well known for its slow convergence rate, especially in dynamic and complicated environments. Instead of adopting deep reinforcement learning structures e.g., double deep QL, deep dueling, in this article, we leverage the fact that there are usually multiple vehicles on the road to speed up the convergence to the optimal solution. To that end, we develop a lightweight yet very effective parallel QL algorithm to quickly obtain the optimal policy by simultaneously learning from various vehicles. Specifically, vehicles on the road can act as active learners to help the system simultaneously collect data. Based on the collected data, the proposed parallel QL algorithm can quickly learn the environment information, e.g., RSSI profile, beam’s location, and blockage, to derive the optimal beam association strategy. The proposed parallel QL algorithm does not only require lower complexity but also converge faster than the latest deep learning-based approaches (e.g., double QL, deep dueling). Moreover, unlike deep reinforcement learning methods (e.g., [18]), our proposed parallel QL framework is proven to always converge to the optimal policy. We show that the high mobility and parallelism of vehicles now become helpful in speeding up the learning process of our underlying algorithm.

Moreover, unlike existing works, e.g., [15][18], in which learning algorithms are deployed at the vehicles with limited resources, in our design, the eNodeB executes the parallel QL algorithm and sends optimal beam association/handover actions to vehicles. As such, the computing complexity is moved to the eNodeB. The simulation results then show that our proposed solution can increase the data rate for each vehicle by up to 47% and reduce the disconnection probability by 29% compared to existing approaches.

I-B Main Contributions

The major contributions of this paper are as follows.

  • Develop an optimal beam association solution for high-mobility mmWave vehicular communication networks using the semi-Markov decision process framework that can effectively deal with the inherent dynamics of the mmWave connection quality and the uncertainty of the environment, e.g., beam’s location, RSSI profile, the velocity of the vehicle, and blockages, in a real-time manner.

  • Develop a lightweight yet very effective parallel QL algorithm to quickly obtain the optimal policy by simultaneously learning from various vehicles to update the global Q-table at the eNodeB. The proposed parallel QL algorithm does not only require lower complexity but also converge faster than advanced deep reinforcement learning approaches (e.g., deep double and deep dueling). Specifically, the algorithm deploys multiple learning processes at the eNodeB, and each learning process is assigned to learn from a vehicle on the considered road to update the global policy.

  • Prove that the proposed parallel QL framework converges with probability one to the optimal policy. Note that deep QL based approaches (e.g., deep double and deep dueling) with advanced deep neural networks are not always guaranteed to converge. We also provide a comprehensive analysis of the convergence time/rate, complexity, and overhead of the proposed framework.

  • Perform extensive simulations to demonstrate the effectiveness of the proposed parallel QL algorithm. By learning from multiple vehicles and exploiting the high mobility of vehicles, our proposed algorithm can achieve the performance close to that of the hypothetical scheme which requires complete environment information in advance.

The rest of paper is organized as follows. The system model is described in Section II. Section III presents the problem formulation based on the SMDP. Then, Section IV describes the conventional QL algorithm and our proposed parallel QL algorithm. After that, we provide the evaluation results in Section V. Finally, the conclusion is highlighted in Section VI.

II System Model

Refer to caption
Figure 1: System model.

Consider a millimeter wave (mmWave) vehicular network, where an LTE eNodeB and a set of NN mmWave base stations (mmBSs) BS={BS1,,BSn,,BSN}BS=\{BS_{1},\ldots,BS_{n},\ldots,BS_{N}\} are deployed as shown in Fig. 1. This is an expected network model for 5G and beyond systems [19][20]. All the mmBSs can connect to the eNodeB via backhaul links. Each vehicle is equipped with two communication interfaces: (i) an LTE interface to communicate with the eNodeB and (ii) an mmWave interface to communicate with an mmBS [15]. The Friss free-space equation reveals that with given transmit power and antenna gains, the pathloss increases when the frequency increases [2][34]. For example, increasing the carrier frequency from 2 GHz to 60 GHz 11160 GHz is the common frequency band in automotive communications. results in an increase of 29 dB for the pathloss. The mmWave signals also suffer from obstacles and scattering objects (both static and dynamic). Thus, the path loss model can be formulated as follows [2][34]:

PL(d)[dB]\displaystyle PL(d)[dB] =PL(d0)+10nlog10dd0+ψ\displaystyle=PL(d_{0})+10n\log_{10}\frac{d}{d_{0}}+\psi (1)
=20log104πd0λ+10nlog10dd0+ψ,\displaystyle=20\log_{10}\frac{4\pi d_{0}}{\lambda}+10n\log_{10}\frac{d}{d_{0}}+\psi,

where d0d_{0} is the reference distance, PL(d0)PL(d_{0}) is the free-space loss at the reference distance, nn is the path loss exponent, dd is the distance between the vehicle and the mmBS, λ\lambda is the carrier wavelength, and ψ\psi is the log-normal shadowing loss caused by the absorption of obstacles and scattering objects. Moreover, the log-normal shadowing loss increases with distance between the vehicle and the mmBS as the farther the transmitter and the receiver are the more likely that there are obstructing objects in between them. When the path loss increases, the received power at the mmBS decreases. Thus, when the vehicle leaves a beam or moves to a blockage zone, the received signal strength indicator (RSSI) will drop. As such, the vehicle is not be able to communicate with its connected mmBS (through the beam). To avoid disrupting the service, the vehicle needs to connect to another beam which provides better channel quality. This beam can belong to the current connected mmBS, i.e., beam association, or belong to another mmBS, i.e., handover.

Conventional methods in the literature [27]-[29] usually make a beam association/handover decision based on the current channel information or network state, e.g., SINR or RSSI, where the decision is triggered when the SINR/RSSI are dropped due to blockage objects or mobility. However, these solutions may lead to too frequent handover and hence the associated handover cost/interruption, especially in mmWave networks where the temporal degradation of channel quality frequently occurs due to the intermittent connectivity of mmWave links. In addition, as the mmWave communication quality often deteriorates intermittently and rapidly, these solutions may lead to the ping-pong effect resulting in high outage probability and low system performance [9][30].

In this paper, our learning algorithm can efficiently address these critical problems by learning the environment conditions. In particular, we consider a centralized controller, i.e., the eNodeB, that “learns” from vehicles on the considered road and makes beam association/handover decisions for all vehicles. With the proposed learning algorithm, the eNodeB can learn the RSSI profiles of the beams in the systems. To reduce the number of beam association/handovers, the eNodeB can guide the vehicle to connect to a beam with a “better” RSSI profile (in terms of the long-run average data rate). In addition, the beam association/handover decision can be triggered even when the RSSI level is still good to anticipate the intermittent problem of mmWave links. It is worth noting that the RSSI level can be inferred by the mmBS through the received signals from the vehicle. Moreover, with current standards in ITS systems [35], the mmBS always has the location of its connected vehicles. As such, the eNodeB runs the algorithm to obtain the best beam for the vehicle to connect without adding noticeable overhead to current ITS systems.

We assume that each mmBS BSnBS_{n} has a finite set n={bn,1,,bn,k,,bn,K}\mathcal{B}_{n}=\{b_{n,1},\ldots,b_{n,k},\ldots,b_{n,\mathrm{K}}\} of KK orthogonal beams [15]. Based on the information learned from the vehicle, i.e., location, velocity, and RSSI level, the eNodeB selects beam bn,kb_{n,k} of mmBS BSnBS_{n} to support the communications of the vehicle. In this paper, the velocity of each vehicle is not fixed. When the vehicle is connected to beam bn,kb_{n,k}, it can successfully transmit data with rate rn,kr_{n,k}. Note that rn,kr_{n,k} is a random variable, depending on the RSSI level (i.e., channel quality) of the channel. Consider MM RSSI levels ={0,1,,M1}\mathcal{R}=\{0,1,\ldots,M-1\} which depend on the environmental conditions, e.g., channel conditions (as modeled above) and blockage probability. The higher the RSSI level is, the higher the achievable data rate of the vehicle. We assume that when the vehicle enters blockage zone, the RSSI level drops to 0, and thus the vehicle cannot connect to the mmBS. We then define ωbn,k\omega_{b_{n,k}} is the blockage probability of beam bn,kb_{n,k} with 0ωbn,k10\leq\omega_{b_{n,k}}\leq 1. ωbn,k=1\omega_{b_{n,k}}=1 if there are static blockage objects (e.g., buildings) in the coverage of the beam. Note that the blockage model is not the input of our proposed algorithm, i.e., it is not required to be available in advance. Instead, our proposed algorithm can learn the blockage model by interacting with the environment and observing the immediate reward. Thus, our proposed framework can work with any probabilistic blockage model (i.i.d or non i.i.d). Similar to [15], we assume that rn,kr_{n,k} is varied from 0 to RmaxR_{\max}, where RmaxR_{\max} is the maximum achievable rate, corresponding to the highest RSSI level. Formally, rn,kr_{n,k} can be formulated as in (2).

rn,k={0,with probability ωbn,k,Δ(l),l{{0}},with probability 1ωbn,k,r_{n,k}=\left\{\begin{array}[]{ll}0,&\mbox{with probability $\omega_{b_{n,k}}$},\\ \Delta(l),l\in\{\mathcal{R}\setminus\{0\}\},&\mbox{with probability $1-\omega_{b_{n,k}}$},\end{array}\right. (2)

where Δ(l)\Delta(l) is the rate corresponding to the current RSSI level ll [15]. Note that to capture the fading effect of the channel, Δ(l)\Delta(l) follows a given random distribution.

Note that although the above environment information (e.g., the RSSI profiles and the blockage probability) are required for the modeling/formulation purpose, the proposed parallel QL algorithm below does not require these parameters explicitly as input. Instead, after executing an action, the eNodeB observes the reward, i.e., the actual data rate between the vehicle and the connected mmBS. The reward function (defined below) hence captures the communication channel between vehicles and mmBSs (e.g., the bit error rate and fading). The details of the proposed algorithm are described in the next section.

III Problem Formulation

This work aims to deal with the dynamics of blockages, channel quality, and mobility. However, the conventional MDP is not effective in capturing the dynamics and uncertainty of the system. Thus, we propose to use the SMDP [21]. Different from the MDP, in an SMDP, an action is only taken when an event occurs. An SMDP can be defined as a tuple <tj,𝒮,𝒜,r><t_{j},\mathcal{S},\mathcal{A},r>, where 𝒮\mathcal{S} and 𝒜\mathcal{A} are the state space and the action space of the system, respectively. tjt_{j} defines decision epoch jj-th when an event occurs, and rr is the reward function.

III-A State Space

The state space of each vehicle is represented as the discretized space of RSSI levels, the connected beam, the speed, and the direction of the vehicle. Formally, the state space 𝒮\mathcal{S} is defined as follows:

𝒮\displaystyle\mathcal{S}\triangleq {(l,bn,k,v,d):l{0,,m,,M1},\displaystyle\Bigl{\{}(l,b_{n,k},v,d):l\in\{0,\ldots,m,\ldots,M-1\}, (3)
bn,kn{b0,0},n{1,,N},k{1,,K},\displaystyle b_{n,k}\in\mathcal{B}_{n}\cup\{b_{0,0}\},\forall n\in\{1,\ldots,N\},\forall k\in\{1,\ldots,K\},
v{0,,vmax},d{0,1}},\displaystyle v\in\{0,\ldots,v_{\mathrm{max}}\},d\in\{0,1\}\Bigr{\}},

where ll is the RSSI level of the current vehicle and bn,kb_{n,k} is the current connected beam of the current vehicle. vv is the vehicle’s speed, vmaxv_{\mathrm{max}} is the maximum speed of the vehicle, and dd is the direction of the vehicle (0 and 1 represent two directions of the vehicle). b0,0b_{0,0} is a virtual beam used to capture the case that there is no available beam at a given location of the current vehicle. In this case, the RSSI level of the current vehicle is always 0. Note that due to the inherent dynamics of mmWave link’s quality, the mmWave instantaneous RSSI estimate can be less accurate, providing less useful information to decide the next action. This is one of the key motivations of our machine learning-based framework that can learn the RSSI profiles of beams by interacting with the environment and observing the actual reward (transmission rate) other than just relying on the instantaneous RSSI estimate.

The road is modeled as a one-dimension area 𝒲\mathcal{W}\in\mathbb{R} that is discretized to Wz\left\lfloor\frac{W}{z}\right\rfloor zones, where WW is the length of the considered road, zz is the length of each zone, and .:\lfloor.\rfloor:\mathbb{R}\rightarrow\mathbb{N} is the floor function. When the vehicle at location w𝒲w\in\mathcal{W}, the vehicle is at zone wz\left\lfloor\frac{w}{z}\right\rfloor-th. At the current state s𝒮s\in\mathcal{S}, an event ese_{s} is triggered if a vehicle reaches a new zone. Note that as the vehicle’s speed is not fixed, the time interval between two consecutive epochs varies. To capture that, the SMDP is used in this work, instead of the conventional Markov decision process with identical time slots. Specifically, the eNodeB only takes an action when event ese_{s} occurs. Note that unlike existing works that have the location information as part of the state space, our framework captures the location information via events ese_{s} under the SMDP formulation. This allows us to efficiently deal with dynamics and uncertainties during the learning process, which is very beneficial in our considered system with the dynamics of blockages, channel quality, and mobility.

III-B Action Space

When the vehicle reaches a new zone on the road, i.e., event ese_{s} is triggered (given its current state s𝒮s\in\mathcal{S}), the eNodeB decides if the vehicle needs to associate to a new beam or stay on the current beam. The action space 𝒜s\mathcal{A}_{s} is defined as:

𝒜s{a}={bn,k,b0,0},n,k,\mathcal{A}_{s}\triangleq\{a\}=\{b_{n,k},b_{0,0}\},\forall n,\forall k, (4)

where aa is the action made at state ss. a=bn,ka=b_{n,k} if the eNodeB guides the vehicle to connect to beam bn,kb_{n,k}. This includes the case staying with the current beam. a=b0,0a=b_{0,0} if there is no available beam at the current location.

III-C Immediate Reward

In this paper, we aim to maximize the long-term average data rate of the system. As mentioned, at decision epoch tt, if an action is taken so that the vehicle connects to beam bn,kb_{n,k}, it can communicate with a rate of 0rn,kRmax0\leq r_{n,k}\leq R_{max} corresponding to the current RSSI level. The resulting data that the vehicle receives from the mmBS is calculated as jn,ktrn,kj_{n,k}^{t}r_{n,k}, where jn,ktj_{n,k}^{t} is the connection time between two consecutive decision epochs (during which the vehicle can communicate with the mmBS through beam bn,kb_{n,k}). As the algorithm observes the reward at the end of each decision epoch, jn,ktj_{n,k}^{t} is the duration from the time that the vehicle enters the current zone until it leaves to enter the next zone. jn,ktj_{n,k}^{t} hence depends on the speed of the vehicle vtv_{t} at the epoch tt. Practically, the speed vtv_{t} can change from one to another epoch or even during the time jn,ktj_{n,k}^{t}. However, without loss of generality, we assume that the time jn,ktj_{n,k}^{t} between two consecutive epochs is small enough (e.g., by setting the length per zone zz as small as necessary) so that the vehicle’s RSSI level and speed remain unchanged. Thus, connection time jn,ktj_{n,k}^{t} can be calculated as zvt\frac{z}{v_{t}}.

In addition, the service may be interrupted during the handover/beam-switching, denoted as hh, i.e., the time it takes for the vehicle to switch to the new mmBS. We assume that the handover time is the same for all the mmBSs. Taking the handover time into account, at state s𝒮s\in\mathcal{S}, the immediate data rate after performing action aa is obtained in (5).

r(st,at)={(jn,kth)rn,k,if at=bn,k and the vehiclehandovers to a new mmBS,jn,ktrn,k,if at=bn,k and the vehiclestays in the current mmBS,0,otherwise,r(s_{t},a_{t})=\left\{\begin{array}[]{ll}(j_{n,k}^{t}-h)r_{n,k},&\mbox{if $a_{t}=b_{n,k}$ and the vehicle}\\ &\mbox{handovers to a new mmBS},\\ j_{n,k}^{t}r_{n,k},&\mbox{if $a_{t}=b_{n,k}$ and the vehicle}\\ &\mbox{stays in the current mmBS},\\ 0,&\mbox{otherwise},\end{array}\right. (5)

where sts_{t} and ata_{t} are the system state and the action taken at decision epoch tt, respectively and rn,kr_{n,k} is the communication rate when the vehicle connects to beam bn,kb_{n,k} as defined in (2). Note that jn,ktj_{n,k}^{t}, rn,kr_{n,k}, and hh depend on the channel and environment conditions such as blockages, beam’s location, and RSSI profiles. Under our design, they are implicitly learned through interacting with the environment and observing the immediate reward in an online manner.

III-D Optimization Formulation

The decision policy π\pi of the proposed SMDP can be defined as a mapping from the state space to the action space: 𝒮𝒜s\mathcal{S}\rightarrow\mathcal{A}_{s} [21][36]. Thus, with initial state ss, the long-term average data rate is formulated as follows:

π(s)=limT𝔼{t=0Tr(st,π(st))|s0=s}𝔼{t=0Tξt|s0=s},s𝒮,\mathcal{R}_{\pi}(s)=\lim\limits_{T\rightarrow\infty}\frac{\mathbb{E}\{\sum_{t=0}^{T}r(s_{t},\pi(s_{t}))|s_{0}=s\}}{\mathbb{E}\{\sum_{t=0}^{T}\xi_{t}|s_{0}=s\}},\forall s\in\mathcal{S}, (6)

where ξt\xi_{t} is the time interval between the tt-th and (t+1)(t+1)-th decision epochs, π(s)\pi(s) is the action at state ss based on policy π\pi, and rr is the immediate reward after performing an action. In Theorem 1, we will prove that the limit in (6) exits [22].

Theorem 1.

With the number of events in a given time and the number of states in the state space 𝒮\mathcal{S} are finite, we have:

π(s)\displaystyle\mathcal{R}_{\pi}(s) =limT𝔼{t=0Tr(st,π(st))|s0=s}𝔼{t=0Tξt|s0=s}\displaystyle=\lim\limits_{T\rightarrow\infty}\frac{\mathbb{E}\{\sum_{t=0}^{T}r({s}_{t},\pi(s_{t}))|s_{0}=s\}}{\mathbb{E}\{\sum_{t=0}^{T}\xi_{t}|s_{0}=s\}} (7)
=¯πr(s,π(s))¯πy(s,π(s)),s𝒮,\displaystyle=\frac{\overline{\mathcal{L}}_{\pi}r(s,\pi(s))}{\overline{\mathcal{L}}_{\pi}y(s,\pi(s))},\forall s\in\mathcal{S},

where y(s,π(s))y(s,\pi(s)) denotes the expected time interval between two consecutive decision epochs when an action is taken at state ss following policy π\pi. ¯π\overline{\mathcal{L}}_{\pi} is the limiting matrix of the transition probability matrix π\mathcal{L}_{\pi} [21], defined as:

¯π=limT1Tt=0T1πt,\overline{\mathcal{L}}_{\pi}=\lim\limits_{T\rightarrow\infty}\frac{1}{T}\sum_{t=0}^{T-1}\mathcal{L}_{\pi}^{t}, (8)
Proof.

First, we prove the following lemma.

Lemma 1.

The limiting matrix ¯π\overline{\mathcal{L}}_{\pi} of the the transition probability matrix π\mathcal{L}_{\pi} always exists.

Proof.

The proof of Lemma 1 is provided in Appendix A. ∎

As the limiting matrix ¯π\overline{\mathcal{L}}_{\pi} exits (see Lemma 1) and the total probabilities of transiting from a given state to other states equals to 1, i.e., s𝒮¯π(s|s)=1\sum_{s^{\prime}\in\mathcal{S}}\overline{\mathcal{L}}_{\pi}(s^{\prime}|s)=1 , we have:

¯πr(s,π(s))=limT1T+1𝔼{t=0Jr(st,π(st))},s𝒮,\displaystyle\overline{\mathcal{L}}_{\pi}r(s,\pi(s))=\lim\limits_{T\rightarrow\infty}\frac{1}{T+1}\mathbb{E}\Bigl{\{}\sum_{t=0}^{J}r(s_{t},\pi(s_{t}))\Bigr{\}},\forall s\in\mathcal{S}, (9)
¯πy(s,π(s))=limT1T+1𝔼{t=0Tξt},s𝒮.\displaystyle\overline{\mathcal{L}}_{\pi}y(s,\pi(s))=\lim\limits_{T\rightarrow\infty}\frac{1}{T+1}\mathbb{E}\Bigl{\{}\sum_{t=0}^{T}\xi_{t}\Bigr{\}},\forall s\in\mathcal{S}.

Clearly, the long-term average reward in (7) is obtained by taking the ratio of ¯πr(s,π(s))\overline{\mathcal{L}}_{\pi}r(s,\pi(s)) and ¯πy(s,π(s))\overline{\mathcal{L}}_{\pi}y(s,\pi(s)). In addition, the ratio of limits equals to the limit of the ratio. Thus, the long-term average reward in (7) is well defined and exists. ∎

Next, in Theorem 2, we prove that the underlying Markov chain is irreducible, and thus the long-term average date rate (π)\mathcal{R}(\pi) does not depend on the initial state s0s_{0}.

Theorem 2.

For every π\pi, the long-term average date rate (π)\mathcal{R}(\pi) is well defined and does not depend on the initial state, i.e., π(s)=π,s𝒮\mathcal{R}_{\pi}(s)=\mathcal{R}_{\pi},\forall s\in\mathcal{S}.

Proof.

The proof of Theorem 2 is provided in Appendix B. ∎

Then, the long-term average data rate optimization problem can be formulated as follows:

maxπ\displaystyle\max_{\pi} π=¯πr(s,π(s))¯πy(s,π(s))\displaystyle\mathcal{R}_{\pi}=\frac{\overline{\mathcal{L}}_{\pi}r(s,\pi(s))}{\overline{\mathcal{L}}_{\pi}y(s,\pi(s))} (10)
s.t. s𝒮¯π(s|s)=1,s𝒮.\displaystyle\sum_{s^{\prime}\in\mathcal{S}}\overline{\mathcal{L}}_{\pi}(s^{\prime}|s)=1,\forall s\in\mathcal{S}.

Our aim in this work is finding the optimal beam association policy to maximize the long-term average data rate, i.e.,

π=argmaxππ.\pi^{*}=\operatorname*{\arg\!\max}_{\pi}\mathcal{R}_{\pi}. (11)

IV Parallel Reinforcement Learning for Beam Association in High Mobility mmWave Vehicular Networks

In this section, we develop the parallel QL algorithm that obtains the optimal beam association policy much faster than those of the existing reinforcement learning based algorithms (e.g., [18]). For that, we first briefly present the details of the conventional QL algorithm. Related mmWave works in the literature usually adopt the QL and deep QL algorithms to solve their problems. However, with dynamic and complicated system, the QL algorithm usually takes a very long time to obtain the optimal strategy. In addition, the deep QL algorithm (e.g., [18]) requires high performance computing resources and does not always ensure to converge to the optimal policy due to the overestimation of the optimizer. Note that our proposed parallel QL algorithm obtains the optimal beam association policy in an on-line manner, bringing various advantages, compared with the off-line learning approach. First, in mmWave vehicular communication systems, the conditions of the environment may be changed quickly. This is due to the fast and intermittent change in the quality of the mmWave links and the dynamics of blockages. Thus, a good beam/mmBS at the current time may quickly become worse later. For that, if we use the off-line training, the framework may not be able to obtain the “current” optimal policy in time, and thus leading to poor learning performance, especially in dynamic environments like mmWave networks. Under the proposed on-line training, channel conditions can be quickly learned to obtain the optimal beam association policy for vehicles. Second, off-line training requires frequent maintenance and adjustment to adapt to the new conditions of the environment (e.g., new buildings, new bus/transportation schedules). Differently, our on-line training can reduce human intervention and management costs as the algorithm can automatically learn the environment conditions and adjust its optimal policy.

IV-A Q-Learning Approach

This section presents the QL algorithm [26], which enable the eNodeB to obtain the optimal beam association strategy for vehicles without prior environment parameters, e.g., RSSI profiles and blockages. The key idea of the QL algorithm is updating the Q-value function for all state-action pairs stored in a Q-table. At a given system state, the QL algorithm performs an action and observes the immediate reward as well as the next state of the system. Based on these observations, the algorithm can update the Q-value for the current state-action pair based on the Q-value function [26]. As such, the learning process is able to learn from the previous experiences, i.e., current state, action, next state, and immediate reward, to derive the optimal solution [26]. In the following, we present the fundamentals of the Q-value function.

We first define the beam association policy as π:𝒮𝒜\pi:\mathcal{S}\rightarrow\mathcal{A}. In particular, π\pi is a mapping from a given state to its corresponding action. Our aim in this paper is finding the optimal beam association policy π\pi^{*} to optimize the system performance in terms of the average data rate, disconnection probability, and number of handovers. Then, we define 𝒱π(s):𝒮\mathcal{V}^{\pi}(s):\mathcal{S}\rightarrow\mathbb{R} as the expected value function of state s𝒮s\in\mathcal{S} given policy π\pi. 𝒱π(s)\mathcal{V}^{\pi}(s) can be formulated as follows:

𝒱π(s)\displaystyle\mathcal{V}^{\pi}(s) =𝔼π[t=0γtrt(st,at)|s0=s]\displaystyle=\mathbb{E}_{\pi}\Big{[}\sum_{t=0}^{\infty}\gamma^{t}r_{t}(s_{t},a_{t})|s_{0}=s\Big{]} (12)
=𝔼π[rt(st,at)+γ𝒱π(st+1)|s0=s],\displaystyle=\mathbb{E}_{\pi}\Big{[}r_{t}(s_{t},a_{t})+\gamma\mathcal{V}^{\pi}(s_{t+1})|s_{0}=s\Big{]},

where 0γ<10\leq\gamma<1 denotes the discount factor. In particular, γ\gamma represents the effect of the future rewards. The higher the discount factor is, the more important future rewards are. At each state ss, we aim to find the optimal action to derive the optimal beam association policy π\pi^{*}, which is a map from a given state to the optimal action. To do that, the optimal value function for each state has to be obtained as in (13).

𝒱(s)=maxa{𝔼π[rt(st,at)+γ𝒱π(st+1)]},s𝒮.\mathcal{V}^{*}(s)=\max_{a}\Big{\{}\mathbb{E}_{\pi}[r_{t}(s_{t},a_{t})+\gamma\mathcal{V}^{\pi}(s_{t+1})]\Big{\}},\quad\forall s\in\mathcal{S}. (13)

We then denote the optimal Q-function state-action pair (s,a),s𝒮,a𝒜(s,a),\forall s\in\mathcal{S},\forall a\in\mathcal{A} as follows:

𝒬(s,a)rt(st,at)+γ𝔼π[𝒱π(st+1)].\mathcal{Q}^{*}(s,a)\triangleq r_{t}(s_{t},a_{t})+\gamma\mathbb{E}_{\pi}[\mathcal{V}^{\pi}(s_{t+1})]. (14)

Hence, the optimal value function is written as follows:

𝒱(s)=maxa{𝒬(s,a)}.\mathcal{V}^{*}(s)=\max_{a}\{\mathcal{Q}^{*}(s,a)\}. (15)

To solve (15), we can update the Q-function to determine the optimal Q-values of all state-action pairs by using (16[26]:

𝒬t+1(st,at)=𝒬t\displaystyle\mathcal{Q}_{t+1}(s_{t},a_{t})=\mathcal{Q}_{t} (st,at)+τt[rt(st,at)\displaystyle(s_{t},a_{t})+\tau_{t}\Big{[}r_{t}(s_{t},a_{t}) (16)
+γmaxat+1𝒬t(st+1,at+1)𝒬t(st,at)],\displaystyle+\gamma\max_{a_{t+1}}\mathcal{Q}_{t}(s_{t+1},a_{t+1})-\mathcal{Q}_{t}(s_{t},a_{t})\Big{]},

where τt\tau_{t} denotes the learning rate, determining the impact of new experiences to the current Q-value [26]. By updating the Q-value functions of all state-action pairs by using (16), the algorithm can derive the optimal beam association policy.

IV-B Parallel Q-Learning Approach

Refer to caption
Figure 2: Parallel QL based model enables simultaneous learning from multiple vehicles.

Note that the conventional QL algorithm can converge to the optimal beam association policy quickly when the system is simple. However, with the dynamics and uncertainties of the system considered in this work, the QL algorithm may take a very long time to obtain the optimal strategy. This is due to the fact that the QL algorithm require a huge number of training episodes to collect enough data for learning. To speed up the QL algorithm, the deep reinforcement learning algorithm is usually adopted in the literature. Nevertheless, this algorithm requires high performance computing resources and does not ensure to converge to the optimal policy due to the overestimation of the optimizer [22][23]. The convergence rate of the deep reinforcement learning algorithm can be improved by implementing several techniques. For example, by using the deep dueling neural network architecture [24], one can significantly improve the convergence rate and stability of the QL algorithm. However, in our work, we leverage the fact that there are simultaneously multiple vehicles running on the road to design the parallel QL algorithm. The proposed algorithm does not only require lower complexity but also converge faster than the latest deep learning-based approaches. In particular, vehicles running on the road act as active learners which can help the system simultaneously collect data and significantly speed up the learning process as shown in Fig. 2.

To that end, the parallel QL algorithm employs multiple learning processes. Each learning process is assigned for a vehicle running on the road (in the coverage of the eNodeB). Specifically, each learning process ii updates the Q-value function at the global Q-table as follows:

𝒬t+1(sti,ati)=𝒬t\displaystyle\mathcal{Q}_{t+1}(s_{t}^{i},a_{t}^{i})=\mathcal{Q}_{t} (sti,ati)+τt[rti(sti,ati)\displaystyle(s_{t}^{i},a_{t}^{i})+\tau_{t}\Big{[}r_{t}^{i}(s_{t}^{i},a_{t}^{i}) (17)
+γmaxat+1i𝒬t(st+1i,at+1i)𝒬t(sti,ati)],\displaystyle+\gamma\max_{a_{t+1}^{i}}\mathcal{Q}_{t}(s_{t+1}^{i},a_{t+1}^{i})-\mathcal{Q}_{t}(s_{t}^{i},a_{t}^{i})\Big{]},

where 0γ<10\leq\gamma<1 is the discount factor that presents the effect of future rewards [26]. In particular, when γ\gamma is low, e.g., close to 0, the learning process prefers the current reward. Differently, when γ\gamma is high, e.g., close to 1, the long-term reward will be considered. In this work, we set γ\gamma the same for all vehicles, i.e., learners. rti(sti,ati)r_{t}^{i}(s_{t}^{i},a_{t}^{i}) is the immediate reward when vehicle ii performs action atia_{t}^{i} at state stis_{t}^{i} (computed using equation 5 above). τt\tau_{t} is the learning rate at decision epoch tt [26]. Note that the learning rate can be fixed at a constant value or it can be adjusted when running the algorithm. In this paper, the learning rate is fixed during the training process and is the same for all learning processes. At each decision epoch and given a current state, i.e., RSSI level and current connected beam, the current vehicle chooses to connect to a beam (following the current beam association policy sent from the eNodeB) and observes the data rate of the connected beam as well as the next state. Then, these observations are sent to the eNodeB for learning by the corresponding learning process to update the global Q-table (equation 17). Algorithm 1 describes the proposed parallel QL algorithm.

Algorithm 1 Parallel QL Algorithm for Vehicle ii
1:for t=1 to T do
2:     Vehicle ii observes the current state sti𝒮s_{t}^{i}\in\mathcal{S} and execute action ati𝒜a_{t}^{i}\in\mathcal{A} based on the ϵ\epsilon-greedy policy.
3:     Vehicle ii observes the immediate reward rtir_{t}^{i} and new state st+1i𝒮s_{t+1}^{i}\in\mathcal{S}.
4:     Vehicle ii sends transition (sti,ati,rti,st+1i)(s_{t}^{i},a_{t}^{i},r_{t}^{i},s_{t+1}^{i}) to the eNodeB for learning by updating the table entry of 𝒬(sti,ati)\mathcal{Q}(s_{t}^{i},a_{t}^{i}) as as follows:
𝒬t+1(sti,ati)=𝒬t\displaystyle\mathcal{Q}_{t+1}(s_{t}^{i},a_{t}^{i})=\mathcal{Q}_{t} (sti,ati)+τt[rti(sti,ati)\displaystyle(s_{t}^{i},a_{t}^{i})+\tau_{t}\Big{[}r_{t}^{i}(s_{t}^{i},a_{t}^{i}) (18)
+γmaxat+1i𝒬t(st+1i,at+1i)𝒬t(sti,ati)]\displaystyle+\gamma\max_{a_{t+1}^{i}}\mathcal{Q}_{t}(s_{t+1}^{i},a_{t+1}^{i})-\mathcal{Q}_{t}(s_{t}^{i},a_{t}^{i})\Big{]}
5:     Replace stist+1is_{t}^{i}\leftarrow s_{t+1}^{i}.
6:end for

In particular, vehicle ii first observes the current state sti𝒮s_{t}^{i}\in\mathcal{S} and performs action atia_{t}^{i} based on the ϵ\epsilon-greedy policy [32]. Then, the eNodeB selects an action that maximizes the Q-value function with probability 1ϵ1-\epsilon and a random action with probability ϵ\epsilon. Then the eNodeB sends this action to vehicle ii to perform. In this work, we gradually reduce the value of ϵ\epsilon. In other words, the algorithm first chooses random actions and gradually change to the deterministic strategy, i.e., choose an action with the highest Q-value at a given state. To that end, ϵ\epsilon is set at a high value (e.g., 1) when the algorithm starts running. Then, at later iterations, the value of ϵ\epsilon is slowly reduced to a small value (e.g., 0.01). After performing action atia_{t}^{i}, vehicle ii observes immediate data rate rti(sti,ati)r_{t}^{i}(s_{t}^{i},a_{t}^{i}) and next state st+1is_{t+1}^{i}. These observations are then sent to the eNodeB for learning. Note that the learning process of each vehicle is independent from others, and all the learning processes share the same global Q-table. By doing that, the Q-table is updated with more experiences from multiple vehicles running on the road. As such, the convergence rate and convergence time of the parallel QL algorithm will be better than that of the conventional QL algorithm as demonstrated in Section V.

The convergence of our proposed parallel QL algorithm can be always guaranteed. In particular, as the learning processes update the global Q-table simultaneously, the selection of an action of a learning process may be based on a new Q-value that is updated by other learning processes. Specifically, at state ss, a learning process chooses to perform action aa that maximizes the Q-value function. However, before that instance, another learning process may reach state ss, then take action aa^{\prime}, and update the Q-value of action aa^{\prime} to be the highest value. In this case, action aa may still be selected by the ϵ\epsilon-greedy policy mentioned above. In other words, an iteration for exploitation becomes an iteration for exploration. Moreover, multiple learning processes (corresponding to multiple vehicles running on the road), which simultaneously update different non-related states, can be considered as independent processes [33]. As a result, the convergence of the proposed parallel Q-learning algorithm is guaranteed and sped up. Given the above, the convergence of the proposed parallel QL algorithm to the optimal policy is formally stated in Theorem 3.

Theorem 3.

Given the learning processes are serializable and under the conditions of τt\tau_{t} in (19), the parallel QL algorithm is ensured to converge to the optimal policy.

τt[0,1),t=1τt=, and t=1(τt)2<.\tau_{t}\in[0,1),\sum_{t=1}^{\infty}\tau_{t}=\infty,\mbox{ and }\sum_{t=1}^{\infty}(\tau_{t})^{2}<\infty. (19)
Proof.

Theorem 3 is proved in Appendix C. ∎

IV-C Impact of High Mobility on Convergence Time

As mentioned, the proposed parallel QL is particularly useful in vehicular networks where the system can simultaneously learn from experiences of multiple vehicles. Moreover, the high mobility of vehicles is also exploited in collecting new experiences in our proposed framework. In particular, the duration to the next decision epoch can be calculated as follows:

ξt=zvt,\xi_{t}=\frac{z}{v_{t}}, (20)

where vtv_{t} is the speed of the vehicle between the tt-th and (t+1)(t+1)-th decision epochs. After a period of ξt\xi_{t}, the vehicle can collect a new sample of experience at decision epoch (t+1)(t+1)-th, i.e., (st+1i,at+1i,rt+1i,st+2i)(s^{i}_{t+1},a^{i}_{t+1},r^{i}_{t+1},s^{i}_{t+2}). From (20), it is clear that with high speeds, the time to move to the next decision epoch of the vehicle will be short. As such, with a given time period, with higher speeds, the vehicle can collect more experiences. Thus, the parallel QL algorithm can converge to the optimal policy faster. In the simulations below, we can observe that the convergence time reduces from 6,0006,000s to 1,0001,000s when the speed increases from 2 m/s to 9 m/s.

IV-D Complexity and Overhead of Parallel Q-Learning

The proposed parallel QL algorithm is efficient with low computational complexity and memory complexity. As mentioned, the state space of our system includes only the current RSSI level, the current connected beam, and the vehicle’s velocity. For that, in common mmWave vehicular networks setting with a few mmBSs, the number of states is small, and thus the size of the global Q-table is also small. Hence, the algorithm can obtain the optimal beam association strategy quickly as the lookup and update table processes are very fast.

Regarding the computational complexity, our proposed algorithm only performs basic calculations without any complex functions as in the other reinforcement learning algorithms, e.g., deep double QL, deep QL, and deep dueling [22][23]. These algorithms implement deep neural networks to approximate the Q-value function to obtain the optimal policy with complicated mathematical operations, e.g., multiply matrices and gradient descent. As a result, they require longer time to process and higher computing resources compared to our proposed parallel QL algorithm. In addition, in this work, we deploy only one Q-table at the eNodeB to store the Q-values for all state-action pairs instead of implementing a separated Q-tables on each vehicle with limited resources. As a result, the computing complexity is moved to the eNodeB which has sufficient resources to obtain the optimal policy in a short time.

Finally, our proposed parallel QL algorithm incurs minimal communication overhead. In particular, the RSSI level can be inferred by the mmBS through the signals received from the vehicle. Moreover, the mmBS can always know which beam the vehicle connected to. These information are sent to the eNodeB through the backhaul link with high bandwidth. Furthermore, in intelligent transport systems [35], the location information of vehicles is frequently reported to the RSU, i.e., mmBS. Thus, to update event ese_{s} of the SMDP, the eNodeB can collect the information of each vehicle through the mmBSs. Therefore, our proposed solution does not add additional overheads to the current ITS standards.

V Performance Evaluation

V-A Parameter Setting

We consider a road with a length of 10001000 meters in the coverage of an eNodeB. On the considered road, 1010 mmBSs are deployed. Each mmBS is equipped with 1010 orthogonal beams. Each beam is assumed to cover an area (on the considered road) ranging randomly from 2020 meters to 5050 meters. The blocking probability (including both temporary and permanent blockages) of each beam is generated randomly from 0 to 11. We define 1010 RSSI levels for each beam corresponding to 1010 data rates ranging from 0 to Rmax=9R_{\max}=9 Gbps, i.e., rn,k{0,1,,9}r_{n,k}\in\{0,1,\ldots,9\}. Unless otherwise stated, at each decision epoch, a vehicle enters the road with probability λ=0.5\lambda=0.5. The handover time is set at 0.50.5 seconds [9][31]. zz is set at 55 meters. The average vehicle speed is set at 77 m/s (about 25 km/h), which is a typical vehicle urban speed [18]. During a decision epoch, the vehicle’s speed remains unchanged. The average speed is then varied from 11 m/s to 99 m/s (about 4 km/h to 32 km/h) in several scenarios to demonstrate the effectiveness of the proposed solution under high mobility. It is important to note that the proposed parallel QL algorithm can learn without requiring these parameters in advance. Instead, the algorithm will learn them by interacting with the environment. For the proposed parallel QL and QL algorithms, the learning rate and discount factor are set at 0.10.1 and 0.90.9, respectively. Moreover, for the ϵ\epsilon-greedy method, the initial value of ϵ\epsilon is set at 11 and gradually reduced to 0.010.01.

We compare our proposed algorithm with three other methods: (i) MaxRate, (ii) Blockage Aware, and (iii) Upper Bound.

  • MaxRate: This scheme first explores all available beams at the current location. Then, the beam with the highest RSSI level will be selected to connect. Once the MaxRate scheme selects the best beam, it will keep connecting to this beam until the end of the current decision epoch. This scheme is used to show the performance of non-adaptive and greedy solutions. As in mmWave systems the temporal degradation of the channel quality frequently occurs, the best beam at the current time may become worse later. Thus, this scheme results in poor system performance in terms of data rate, the number of handovers, and disconnection probability.

  • Blockage Aware: This scheme is assumed to know the prior knowledge about the blocking probability of all available beams in the current location. Then, the beam with the lowest blocking probability will be chosen for the vehicle to connect. This scheme is used to show the effects of the high mobility. Specifically, when the vehicle speed is low, this scheme will achieve a good system performance in terms of data rate, the number of handovers, and disconnection probability. However, its performance becomes worse when the vehicle speed is high as the high mobility is not considered.

  • Upper Bound: Similar to the Blockage Aware scheme, this scheme is assumed to know the prior knowledge about the blocking probability of all available beams in the current location. However, the handover decision is only made if the data rate achieved is higher than that of staying in the current beam. Note that in practice, this data rate is not available in advance and depended on the channel quality. This scheme is adopted to show the optimistic upper bound of the system performance.

The evaluation metrics are the average data rate, the disconnection probability, and the number of handovers. The average data rate is defined as the data received (in bits) by a vehicle running on the road in a second. The number of handovers is defined as the total number of handovers that a vehicle needs to do when running on the road. The disconnection probability is defined as the average probability that the vehicle cannot communicate with mmBSs.

V-B Simulation Results

Performance Evaluation

We first vary the average speed of vehicles running on the road and evaluate the system performance obtained by the parallel QL algorithm in terms of the average data rate, the disconnection probability, and the number of handovers as shown in Fig. 3. Obviously, the average data rate of the vehicle decreases when the vehicle speed increases as shown in Fig. 3(a). The reason is due to the effects of mobility on the handover. In particular, in the case that the vehicle chooses to handover to a new beam and its speed is high, the vehicle may move to another beam before the handover is finished. Moreover, at high speeds, the estimated RSSI level can be less accurate. This results in lower data rates. It is worth noting that by learning the environment parameters, our proposed solution achieves higher data rates than that of the MaxRate scheme and the achieved rate is close to the Upper Bound scheme. Moreover, when the vehicle speed is high (i.e., 7\geq 7 m/s), our proposed solution achieves a higher data rate than that of the Blockage Aware scheme. The reason is that the Blockage Aware scheme selects the beam with the lowest blocking probability at a given location without considering the vehicle speed. However, at high speeds, the beam with the lowest blocking probability may not be the best choice as the vehicle may move out of the beam before completing the handover process, and thus resulting in low data rates. As can be seen in Fig. 3(b), when the speed increases from 1 m/s to 6 m/s, our proposed solution chooses to reduce the number of handovers to avoid the negative effect of mobility. In contrast, other solutions with fixed policies cannot learn this information, and thus they do not reduce the number of handovers, except the Upper Bound scheme which is assumed to know the actual data rate before connecting to a beam. It is worth noting that, when the speed is higher than 77 m/s (about 25 km/h), the number of handovers slightly increases for all schemes. This is because at high speeds, the vehicle moves out the coverage of a beam before finishing the handover to connect to this beam, thereby it needs to do the handover again. It is important to note that, at several locations, the only option is to handover to a new beam. Thus, the number of handovers slightly increases for our proposed solution in this case. Finally, as shown in Fig. 3(c), the average disconnection probabilities of all schemes increase when the vehicle’s speed increases. The reason is due to the effects of the high mobility on the handover process. Note that by using the learning algorithm, our proposed solution achieves lower disconnection probability than that of the MaxRate scheme and the probability is very close to that of the Upper Bound scheme, especially at high speeds.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 3: (a) Average data rate (Gbit/s), (b) average number of handovers, and (c) average disconnection probability vs. average speed of vehicles.
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 4: (a) Average data rate (Gbit/s), (b) average number of handovers, and (c) average disconnection probability vs. hh.

Next, we fix the average speed of vehicles at 77 m/s (about 25 km/h), which is a typical vehicle urban speed [18] and vary the time for the handover to show the average data rate, number of handovers, and disconnection probability obtained by the proposed parallel QL algorithm as shown in Fig. 4. Clearly, when the time for the handover increases, the disconnection probability increases, and thus the average data rates of all solution decrease as shown in Fig. 4(a) and Fig. 4(c). Again, the proposed solution possesses better performance in terms of data rate and disconnection probability than those of the MaxRate scheme and close to that of the Upper Bound solution. The reason is that our proposed solution can learn and minimize the number of handovers when the time for the handover increases. In particular, as shown in Fig. 4(b), when the time for handover increases from 0.10.1 seconds to 0.40.4 seconds, the number of handovers of the proposed solution decreases as our learning algorithm adapts its policy to minimize the number of the handovers and thus maximize the data rate for the vehicle (when the environment parameters are not available in advance). However, when the time for the handovers is too long (i.e., 0.5\geq 0.5), the number of handovers increases. Similar to the previous scenario, when the handover is too long, the vehicle may move out of the target beam, and thus it needs to do the handover again. It is worth noting that at high speeds, our proposed solution performs betters than the Blockage Aware scheme with prior knowledge about the beam profiles. The reason is that the Blockage Aware scheme selects a beam based on its blocking probability without considering the handover process.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 5: (a) Average data rate (Gbit/s), (b) average number of handovers, and (c) average disconnection probability vs. λ\lambda.

Finally, in Fig. 5, we vary the probability that a vehicle enters the road at each decision epoch to evaluate the performance of the proposed solution. Similar to other scenarios, by learning the environment parameters, i.e., blocking probability, vehicle’s speed, and handover time, our proposed solution achieves better performance than that of the MaxRate and Blockage Aware schemes in terms of data rate, number of handovers, and disconnection probability.

Convergence

Refer to caption
Figure 6: Convergence rates of QL and parallel QL algorithms.

In Fig. 6, we evaluate the convergence rates of the proposed parallel QL and the QL algorithms. Obviously, the parallel QL algorithm can obtain the optimal beam association policy within 6,0006,000 iterations while the QL algorithm cannot converge to the optimal beam association policy after 10,00010,000 iterations. This result confirms the analysis in Section IV-B. Specifically, by learning from multiple vehicles on the road at the same time, the proposed parallel algorithm has more experiences to learn and quickly converge to the optimal policy.

Refer to caption
Figure 7: Convergence rates of the parallel QL algorithm with different numbers of learners.

Next, in Fig. 7, we compare the convergence rates of the parallel QL with different numbers of learners. Clearly, the higher number of learners results in better performance. In particular, when the parallel QL runs with only 2 learners, the performance is the worst. When the number of learners increases to 10, the performance of the algorithm is improved. Finally, in the case that we do not limit the number of learners (i.e., learning from all the vehicles running on the road), the algorithm achieves the best performance and quickly convergences to the optimal beam association policy. This implies that by leveraging the fact that there are often numerous vehicles on the road, our proposed parallel QL algorithm can significantly improve the system performance compared to conventional methods.

Refer to caption
(a)
Refer to caption
(b)
Figure 8: Convergence rates of the QL and parallel QL algorithms when the average speed of vehicles is (a) 2 m/s and (b) 9 m/s.

Next, we investigate the convergence time of the proposed parallel QL and the QL algorithms when the average speed of vehicles is low (2 m/s) and high (9 m/s) as shown in Fig. 8(a) and Fig. 8(b), respectively. As discussed in Section IV-C, increasing the speed of vehicles leads to better convergence time of the algorithm as vehicles can collect more experiences for the learning process. As shown in Fig. 8(a), when the speed of vehicles is 2 m/s, the parallel QL algorithm requires at least 14,00014,000 seconds to obtain the optimal beam association/handover policy. In contrast, when the speed of vehicles is 9 m/s, the algorithm can obtain the optimal association policy within 3,0003,000 seconds as shown in Fig. 8(b). Meanwhile, the QL algorithm still cannot converge to the optimal policy after 20,00020,000 seconds and 10,00010,000 seconds when the vehicle speed is 2 m/s and 9 m/s, respectively. Note that the average data rate (after obtaining the optimal beam association strategy) achieved by the proposed solution in the case the speed of vehicles is 9 m/s lower than that of the case when the speed of vehicles is 2 m/s. As mentioned, this is stemmed from the effects of mobility on the handover process as discussed in the above section.

Refer to caption
Figure 9: Convergence rates of parallel QL and deep dueling algorithms.

Finally, in Fig. 9, we compare the convergence rates of the proposed algorithm and the latest advance in deep reinforcement learning, i.e., deep dueling algorithm. In particular, the deep dueling reinforcement learning algorithm implements two flows of hidden layers to separately estimate the advantage and value functions [22][23]. Recent works demonstrated that the deep dueling algorithm is superior to other deep reinforcement learning algorithms, e.g., deep QL and deep double QL [22][23]. As shown in Fig. 9, our proposed solution can obtain the optimal beam association strategy within 6,0006,000 iterations while the deep dueling cannot converge to the optimal solution after 10,00010,000 iterations. The reason is that our proposed parallel QL algorithm can learn from multiple vehicles running on the road simultaneously. In contrast, the deep dueling algorithm only learns from a single vehicle at a time resulting in poor performance.

VI Conclusion

In this paper, we have developed an optimal beam association framework for high mobility mmWave vehicular networks, aiming to maximize the system performance in terms of average data rates, number of handovers, and disconnection probability of vehicles. The proposed parallel QL algorithm leverages the inherent feature of vehicular networks that there are usually multiple vehicles on the road. By collecting experiences/samples simultaneously from all vehicles, the algorithm converges to the optimal policy much faster than the conventional QL and even its latest deep reinforcement learning framework. Extensive simulations have proved that our proposed parallel QL algorithm can increase the average data rate by 47% and reduce the disconnection probability by 29% compared to the conventional solution. In addition, by learning the RSSI profiles of beams and blockages on the road, our proposed solution can achieve the performance close to that of the hypothetical scheme which requires complete environment information in advance. We also observed that the high mobility of the vehicles was actually helpful in speeding up the convergence of the algorithm to the optimal policy.

Appendix A The proof of Lemma 1

First, we define a sequence of matrices as {An:n0}\{A_{n}:n\geq 0\}. If limnAn(s|s)=(s|s),(s,s)𝒮×𝒮\lim\limits_{n\rightarrow\infty}A_{n}(s^{\prime}|s)=(s^{\prime}|s),\forall(s,s^{\prime})\in\mathcal{S}\times\mathcal{S}, we have limnAn=A\lim\limits_{n\rightarrow\infty}A_{n}=A. Now, we define the Cesaro limit (denoted by C-lim) [21] of the sequence as follows:

C-limN=limNA0+A1++An++AN1N.C\mbox{-}\!\!\lim\limits_{\!\!\!N\rightarrow\infty}=\lim\limits_{N\rightarrow\infty}\frac{A_{0}+A_{1}+\ldots+A_{n}+\ldots+A_{N-1}}{N}. (21)

Thus, AA is the Cesaro limit (of order one) of {An:n0}\{A_{n}:n\geq 0\} if

limN1Nn=0N1An=A.\lim\limits_{N\rightarrow\infty}\frac{1}{N}\sum_{n=0}^{N-1}A_{n}=A. (22)

In a short form, we have C-limNAN=AC\mbox{-}\!\!\lim\limits_{\!\!\!N\rightarrow\infty}A_{N}=A. The limiting matrix ¯\overline{\mathcal{L}} is then formulated as follows:

¯=C-limNN.\overline{\mathcal{L}}=C\mbox{-}\!\!\lim\limits_{\!\!\!N\rightarrow\infty}\mathcal{L}^{N}. (23)

Let’s denote l¯(s|s)\overline{l}(s^{\prime}|s) as the (s|s)(s^{\prime}|s)-th element of ¯\overline{\mathcal{L}}. Thus, for each ss and ss^{\prime}, we have the following:

l¯(s|s)=limN1Nn=1Nln1(s|s),\overline{l}(s^{\prime}|s)=\lim_{N\rightarrow\infty}\frac{1}{N}\sum_{n=1}^{N}l^{n-1}(s^{\prime}|s), (24)

where l0(s|s)l^{0}(s^{\prime}|s) denotes a element of an 𝒮×𝒮\mathcal{S}\times\mathcal{S} identity matrix, and ln1l^{n-1} is a component of n1\mathcal{L}^{n-1}. Given \mathcal{L} is aperiodic, limNN\lim_{N\rightarrow\infty}\mathcal{L}^{N} equals to ¯\overline{\mathcal{L}}. Thus, the limiting matrix exists.

Appendix B The proof of Theorem 2

In this proof, we first show that the underlying Markov chain is irreducible. In particular, we will prove that the learning process can move from a given state to any states after a finite number of steps. As mentioned, the system state space 𝒮\mathcal{S} is the combination of the RSSI level, the connected beam, and the velocity of the current vehicle. At state s=(l,bn,k,v,d)s=(l,b_{n,k},v,d), if the vehicle to connect to beam bn,kb_{n^{\prime},k^{\prime}} and the RSSI level when connect to this beam is ll^{\prime}, the system moves to state s=(l,bn,k,v,d)s^{\prime}=(l^{\prime},b_{n^{\prime},k^{\prime}},v,d). The new RSSI level ll^{\prime} can be any of levels in \mathcal{R} as the RSSI level depends on the environmental conditions, e.g., channel conditions, and the blockage probability. In addition, the vehicle can be able to connect to all beams when it is moving on the road. When the vehicle moves out of the considered road, the system will wait for a new vehicle enters the road and move to a new state. Moreover, the velocity (i.e., speed and direction) of the vehicle is not fixed. Thus, from a given state ss, the system can move to any other state s𝒮s^{\prime}\in\mathcal{S} after a finite number of steps. In other words, the state space 𝒮\mathcal{S} (which is the combination of the RSSI level, the connected beam, and the velocity of the current vehicle) contains only one communicating class, and the underlying Markov chain is irreducible. As such, the long-term average date rate (π)\mathcal{R}(\pi) does not depend on the initial state and is well defined π\forall\pi [25]. Thus, the algorithm can converge to the optimal association policy regardless of the initial system state.

Appendix C The proof of Theorem 3

In this proof, we show that the proposed parallel QL algorithm is ensured to converge to the optimal policy, i.e., 𝒬t(s,a)𝒬(s,a)\mathcal{Q}_{t}(s,a)\rightarrow\mathcal{Q}^{*}(s,a) as tt\rightarrow\infty. As mentioned in Section IV-B, the learning processes in our proposed algorithm are serializable. Thus, the convergence proof of the parallel QL is similar to that of the QL algorithm.

The key idea of this proof is using the action-replay process (ARP) (i.e., an artificial controlled Markov decision process) [26]. This action-relay process is defined based on the episode sequence and the learning rate. First, we denote {s,t}\{\langle s,t\rangle\} as the state space of the ARP [26]. Here, ss is a state in the actual process, t1t\geq 1 denotes the ARP’s level. In addition, the action space of the ARP is denoted as {a}\{a\} in which aa is a action in the actual process. Next, at state s,t\langle s,t\rangle, if action aa is chosen, the state transition consequence and the stochastic reward of the ARP can be formulated as follows:

𝐢={argmaxi{tit},if (s,a) has been taken beforedecision epoch t,0,otherwise,\mathbf{i_{*}}=\left\{\begin{array}[]{ll}\operatorname*{\arg\!\max}_{i}\{t^{i}\leq t\},&\mbox{if $(s,a)$ has been taken before}\\ &\mbox{decision epoch t},\\ 0,&\mbox{otherwise},\end{array}\right. (25)

where tit^{i} represents the ithi^{th} time when performing action aa given state ss. As such, tit^{i_{*}} is the last time at which action aa is taken at state ss in the real process before decision epoch tt. The reward equals to 𝒬0(s,a)\mathcal{Q}_{0}(s,a) if i=0i_{*}=0. Moreover, in this case, the action-replay process is absorbed. Otherwise, we denote the index of the decision epoch which is taken from the existing samples from the real process as follows:

𝐢𝐞={i,with probability τti,i1,with probability (1τti)τti1,i2,with probability (1τti)(1τti1)τti2,0,with probability i=1i(1τti),\mathbf{i_{e}}=\left\{\begin{array}[]{ll}i_{*},&\mbox{with probability $\tau_{t^{i_{*}}}$},\\ i_{*}-1,&\mbox{with probability $(1-\tau_{t^{i_{*}}})\tau_{t^{i_{*}-1}}$},\\ i_{*}-2,&\mbox{with probability $(1-\tau_{t^{i_{*}}})(1-\tau_{t^{i_{*}-1}})\tau_{t^{i_{*}-2}}$},\\ &\vdots\\ 0,&\mbox{with probability $\prod_{i=1}^{i_{*}}(1-\tau_{t^{i}})$},\end{array}\right. (26)

Similar as above, when ie=0i_{e}=0, the reward is 𝒬0(s,a)\mathcal{Q}_{0}(s,a) and the process is absorbed. Otherwise, taking ie#0i_{e}\#0 results reward rtier_{t^{i_{e}}} and a state transition to stie,tie1\langle{s^{\prime}}_{t^{i_{e}}},t^{i_{e}}-1\rangle.

Putting the above and Lemma B in [26] together, we have 𝒬t(s,a)𝒬ARP(s,t,a),a,s,\mathcal{Q}_{t}(s,a)\rightarrow\mathcal{Q}^{*}_{ARP}(\langle s,t\rangle,a),\forall a,s, and t0t\geq 0, in which 𝒬t(s,a)\mathcal{Q}_{t}(s,a) is the optimal action values of the ARP with state s,t\langle s,t\rangle and action aa [26, Lemma A]. Let’s denote rr^{*} as the bound of the reward, and thus r|rt|,tr*\geq|r_{t}|,\forall t. With loss of generality, assuming that 𝒬t(s,a)<r(1γ)\mathcal{Q}_{t}(s,a)<\frac{r^{*}}{(1-\gamma)} with r1r^{*}\geq 1 [26]. Thus, with χ>0\chi>0, we can find ξ\xi so that γξr1γ<χ6.\gamma^{\xi}\frac{r*}{1-\gamma}<\frac{\chi}{6}.

By using Lemma B.4 in [26], the comparison of between the value of performing a1,,aξa_{1},\ldots,a_{\xi} in the real process, i.e., 𝒬¯(s,a1,,aξ)\bar{\mathcal{Q}}(s,a_{1},\ldots,a_{\xi}), with that of taking these actions in the ARP, i.e., 𝒬¯ARP(s,t,a1,,aξ)\bar{\mathcal{Q}}_{ARP}(\langle s,t\rangle,a_{1},\ldots,a_{\xi}), is formulated as follows:

|𝒬¯ARP(<s,t>,a1,,aξ)𝒬¯(s,a1,,aξ)|<\displaystyle|\bar{\mathcal{Q}}_{ARP}(<s,t>,a_{1},\ldots,a_{\xi})-\bar{\mathcal{Q}}(s,a_{1},\ldots,a_{\xi})|< (27)
χ(1γ)6ξr2ξr1γ+2χ3ξ(ξ+1)ξ(ξ+1)2=2χ3.\displaystyle\frac{\chi(1-\gamma)}{6\xi r*}\frac{2\xi r*}{1-\gamma}+\frac{2\chi}{3\xi(\xi+1)}\frac{\xi(\xi+1)}{2}=\frac{2\chi}{3}.

Based on Lemma B.4 in [26], we can say that taking only ξ\xi actions results in a small different of less than χ6\frac{\chi}{6} for both the real process and the action-replay process. In addition, we can apply (27) to an set of actions in both the action-replay and the real processes. As such, 𝒬ARP(s,t,a)𝒬(s,a)|<χ\mathcal{Q}^{*}_{ARP}(\langle s,t\rangle,a)-\mathcal{Q}^{*}(s,a)|<\chi. Thus, 𝒬t(s,a)𝒬(s,a)\mathcal{Q}_{t}(s,a)\rightarrow\mathcal{Q}^{*}(s,a) when tt\rightarrow\infty with probability 1.

References

  • [1] N. V. Huynh, D. N. Nguyen, D. T. Hoang, and E. Dutkiewicz, “Optimal Beam Association in mmWave Vehicular Networks with Parallel Reinforcement Learning”, to be presented in IEEE GLOBECOM, Taipei, Taiwan, Dec. 7-11, 2020.
  • [2] V. Va, T. Shimizu, G. Bansal, and R. W. Heath Jr, “Millimeter wave vehicular communications: A survey,” Foundations and Trends in Networking, vol. 10, no. 1, pp. 1-113, Jun. 2016.
  • [3] D. Yan, H. Yi, D. He, K. Guan, B. Ai, Z. Zhong, J. Kim, and H. Chung, “Channel Characterization for Satellite Link and Terrestrial Link of Vehicular Communication in the mmWave Band,” IEEE Access, vol. 7, pp. 173559-173570, Nov. 2019.
  • [4] A. D. Angelica, “Google’s self-driving car gathers nearly 1 GB/sec”, [Online]. Available: http://www.kurzweilai.net/googles-self-driving-cargathers-nearly-1-gbsec.
  • [5] X. Wang, L. Kong, F. Kong, F. Qiu, M. Xia, S. Arnon, and G. Chen, “Millimeter wave communication: A comprehensive survey,” IEEE Communications Surveys & Tutorials, vol. 20, no. 3, pp. 1616-1653, Jun. 2018.
  • [6] F. B. Mismar, B. L. Evans and A. Alkhateeb, “Deep Reinforcement Learning for 5G Networks: Joint Beamforming, Power Control, and Interference Coordination,” IEEE Transactions on Communications, vol. 68, no. 3, pp. 1581-1592, Mar. 2020.
  • [7] O. Semiari, W. Saad, and M. Bennis, “Joint millimeter wave and microwave resources allocation in cellular networks with dual-mode base stations,” IEEE Transactions on Wireless Communications, vol. 16, no. 7, pp. 4802-4816, Jul. 2017.
  • [8] M. Polese, M. Giordani, M. Mezzavilla, S. Rangan, and M. Zorzi, “Improved handover through dual connectivity in 5G mmWave mobile networks,” IEEE Journal on Selected Areas in Communications, vol. 35, no. 9, pp. 2069-2084, Sept. 2017.
  • [9] S. Zang, W. Bao, P. L. Yeoh, B. Vucetic, and Y. Li, “Managing Vertical Handovers in Millimeter Wave Heterogeneous Networks,” IEEE Transactions on Communications, vol. 67, no. 2, pp. 1629-1644, Feb. 2019.
  • [10] S. Chiu, N. Ronquillo, and T. Javidi, “Active Learning and CSI Acquisition for mmWave Initial Alignment,” IEEE Journal on Selected Areas in Communications, vol. 37, no. 11, pp. 2474-2489, Nov. 2019, doi: 10.1109/JSAC.2019.2933967.
  • [11] F. Sohrabi and W. Yu, “Hybrid analog and digital beamforming for mmWave OFDM large-scale antenna arrays,” IEEE Journal on Selected Areas in Communications, vol. 35, no. 7, pp. 1432-1443, Jul. 2017.
  • [12] N. Garcia, H. Wymeersch, E. G. Strom, and D. Slock, “Location-aided mm-wave channel estimation for vehicular communication,” IEEE 17th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Edinburgh, UK, Jul. 2016.
  • [13] V. Va, T. Shimizu, G. Bansal, and R. W. Heath, “Beam design for beam switching based millimeter wave vehicle-to-infrastructure communications,” IEEE International Conference on Communications (ICC), Kuala Lumpur, Malaysia, Jul. 2016.
  • [14] M. Alrabeiah and A. Alkhateeb, “Deep learning for mmWave beam and blockage prediction using Sub-6 GHz channels,” IEEE Transactions on Communications, vol. 68, no. 9, pp. 5504-5518, Sept. 2020.
  • [15] G. H. Sim, S. Klos, A. Asadi, A. Klein, and M. Hollick, “An online context-aware machine learning algorithm for 5G mmWave vehicular communications,” IEEE/ACM Transactions on Networking, vol. 26, no. 6, pp. 2487-2500, Dec. 2018.
  • [16] M. Scalabrin, N. Michelusi, and M. Rossi, “Beam Training and Data Transmission Optimization in Millimeter-Wave Vehicular Networks,” IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates, 2018.
  • [17] M. Hussain, M. Scalabrin, M. Rossi, and N. Michelusi, “Adaptive Millimeter-Wave Communications Exploiting Mobility and Blockage Dynamics,” IEEE International Conference on Communications (ICC), Dublin, Ireland, 2020.
  • [18] H. Khan, A. Elgabli, S. Samarakoon, M. Bennis, and C. S. Hong, “Reinforcement Learning Based Vehicle-cell Association Algorithm for Highly Mobile Millimeter Wave Communication,” IEEE Transactions on Cognitive Communications and Networking, Early Access, 2019.
  • [19] M. Giordani, M. Mezzavilla, S. Rangan, and M. Zorzi, “Multi-connectivity in 5G mmWave cellular networks,” in Proc. IFIP Med-Hoc-Net, Jun. 2016, pp. 1–7.
  • [20] H. Shokri-Ghadikolaei, F. Boccardi, C. Fischione, G. Fodor, and M. Zorzi,“Spectrum sharing in mmWave cellular networks via cell association, coordination, and beamforming,” IEEE Journal on Selected Areas in Communications, vol. 34, no. 11, pp. 2902–2917, Nov. 2016.
  • [21] M. Puterman, Markov Decision Processes: Discrete Stochastic Dynamic Programming, Hoboken, NJ: Wiley, 1994.
  • [22] N. V. Huynh, D. T. Hoang, D. N. Nguyen, and E. Dutkiewicz, “Optimal and fast real-time resource slicing with deep dueling neural networks”, IEEE Journal on Selected Areas in Communications, vol. 37, no. 6, pp. 1455-1470, Jun. 2019.
  • [23] N. V. Huynh, D. N. Nguyen, D. T. Hoang, and E. Dutkiewicz, “Jam Me If You Can: Defeating Jammer with Deep Dueling Neural Network Architecture and Ambient Backscattering Augmented Communications”, IEEE Journal on Selected Areas in Communications, vol. 37, no. 11, pp. 2603-2620, Nov. 2019.
  • [24] Z. Wang, T. Schaul, M. Hessel, H. V. Hasselt, M. Lanctot, and N. D. Freitas, “Dueling Network Architectures for Deep Reinforcement Learning,” [Online]. Available: arXiv:1511.06581.
  • [25] J. Filar and K. Vrieze, Competitive Markov Decision Processes. Springer Press, 1997.
  • [26] C. J. C. H. Watkins and P. Dayan, “Q-learning,” Mach. Learn., vol. 8, no. 3–4, pp. 279–292, 1992.
  • [27] E. S. Navarro, Y. Lin, and V. Wong, “An MDP-based vertical handoff decision algorithm for heterogeneous wireless networks,” IEEE Transactions on Vehicular Technology, vol. 57, no. 2, pp. 1243-1254, Mar. 2008.
  • [28] M. Mezzavilla, S. Goyal, S. Panwar, S. Rangan, and M. Zorzi, “An MDP model for optimal handover decisions in mmWave cellular networks,” in Proc. EUCNC 2016, Athens, Greece, Jun. 2016, pp. 100-105.
  • [29] H. Elshaer, M. N. Kulkarni, F. Boccardi, J. G. Andrews, and M. Dohler, “Downlink and uplink cell association with traditional macrocells and millimeter wave small cells,” IEEE Transactions on Wireless Communications, vol. 15, no. 9, pp. 6244-6258, Sep. 2016.
  • [30] Y. Sun, G. Feng, S. Qin, Y.-C. Liang, and T.-S. P. Yum, “The smart handoff policy for millimeter wave heterogeneous cellular networks,” IEEE Transactions on Mobile Computing, vol. 17, no. 6, pp. 1456-1468, Jun. 2018.
  • [31] H. S. Park, Y. Lee, T. J. Kim, B. C. Kim, and J. Y. Lee, “Handover mechanism in NR for ultra-reliable low-latency communications,” IEEE Network, vol. 32, no. 2, pp. 41-47, Mar. 2018.
  • [32] R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction. Cambridge, MA, USA: MIT Press, 1998.
  • [33] M. Camelo, J. Famaey, and S. Latre, “A scalable parallel Q-learning algorithm for resource constrained decentralized computing environments,” Proc. Workshop on Machine Learning in HPC Environments (MLHPC), Salt Lake City, UT, USA, Nov. 2016, pp. 27-35.
  • [34] F. Jameel, S. Wyne, S. J. Nawaz, and Z. Chang, “Propagation channels for mmWave vehicular communications: State-of-the-art and future research directions,” IEEE Wireless Communications, vol. 26, no. 1, pp. 144-150, Feb. 2019.
  • [35] ETSI TR 102 638 V1.1.1 Intelligent Transport Systems (ITS); Vehicular Communications; Basic Set of Applications; Definitions.
  • [36] Q. Li, L. Zhao, J. Gao, H. Liang, L. Zhao, and X. Tang, “SMDP-based coordinated virtual machine allocations in cloud-fog computing systems,” IEEE Internet of Things Journal, vol. 5, no. 3, pp. 1977-1988, Jun. 2018.