This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Transfer Generative Adversarial Networks (T-GAN)-based Terahertz Channel Modeling

Zhengdong Hu, Yuanbo Li, and Chong Han Terahertz Wireless Communications (TWC) Laboratory, Shanghai Jiao Tong University, China.
Email: {huzhengdong, yuanbo.li, chong.han}@sjtu.edu.cn
Abstract

Terahertz (THz) communications are envisioned as a promising technology for 6G and beyond wireless systems, providing ultra-broad bandwidth and thus Terabit-per-second (Tbps) data rates. However, as foundation of designing THz communications, channel modeling and characterization are fundamental to scrutinize the potential of the new spectrum. Relied on physical measurements, traditional statistical channel modeling methods suffer from the problem of low accuracy with the assumed certain distributions and empirical parameters. Moreover, it is time-consuming and expensive to acquire extensive channel measurement in the THz band. In this paper, a transfer generative adversarial network (T-GAN) based modeling method is proposed in the THz band, which exploits the advantage of GAN in modeling the complex distribution, and the benefit of transfer learning in transferring the knowledge from a source task to improve generalization about the target task with limited training data. Specifically, to start with, the proposed GAN is pre-trained using the simulated dataset, generated by the standard channel model from 3rd generation partnerships project (3GPP). Furthermore, by transferring the knowledge and fine-tuning the pre-trained GAN, the T-GAN is developed by using the THz measured dataset with a small amount. Experimental results reveal that the distribution of PDPs generated by the proposed T-GAN method shows good agreement with measurement. Moreover, T-GAN achieves good performance in channel modeling, with 9 dB improved root-mean-square error (RMSE) and higher Structure Similarity Index Measure (SSIM), compared with traditional 3GPP method.

I Introduction

With the exponential growth of the number of interconnected devices, the sixth generation (6G) is expected to achieve intelligent connections of everything, anywhere, anytime [1], which demands Tbit/s wireless data rates. To fulfill the demand, Terahertz (THz) communications gain increasing attention as a vital technology of 6G systems, thanks to the ultra-broad bandwidth ranging from tens of GHz to hundreds of GHz [2]. The THz band is promising to address the spectrum scarcity and capacity limitations of current wireless systems, and realize long-awaited applications, extending from wireless cognition, localization/positioning, to integrated sensing and communication [3].

To design reliable THz wireless systems, one fundamental challenge lies in developing an accurate channel model to portray the propagation phenomena. Due to the high frequencies, new characteristics occur in the THz band, such as frequency-selective absorption loss and rough-surface scattering. Attribute to these new characteristics, THz channel modeling is required to capture these characteristics. However, traditional statistical channel modeling methods suffer from the problem of low accuracy with the assumed certain distributions and empirical parameters. For example, a geometric based stochastic channel model (GSCM) assumes that the positions of scatters follow certain statistical distributions, such as the uniform distribution within a circle around the transmitters and receivers [4]. However, the positions of scatters are hard to characterize by certain statistical distributions, making the GSCM not accurate for utilization in the THz band. Moreover, it is time-consuming and costly to acquire extensive channel measurement for THz channel modeling. To this end, an accurate channel modeling method with limited measurement data for the THz band is needed.

Recently, deep learning (DL) is popular and widely applied in wireless communications, such as channel estimation [5, 6] and channel state information (CSI) feedback [7]. Among different kinds of DL methods, the generative adversarial network (GAN) has the advantage of modeling complex distribution accurately without any statistical assumptions, based on which GAN can be utilized to develop channel models. The authors in [8] train GAN to approximate the probability distribution functions (PDFs) of stochastic channel response. In [9], GAN is applied to generate synthetic channel samples close to the distribution of real channel samples. The researchers in [10] model the channel with GAN through channel input-output measurements. In [11], a model-driven GAN-based channel modeling method is developed in intelligent reflecting surface (IRS) aided communication system. These methods achieve good performance in modeling the channel and prove high consistency between the target channel distribution and the generated channel distribution. However, the GAN based channel modeling method has not been exploited in the THz band. Moreover, it is a challenge to train GAN for channel modeling with the scarce THz channel measurement dataset.

In this paper, a transfer GAN (T-GAN)-based THz channel modeling method is proposed, which can learn the distribution of power delay profile (PDP) of the THz channel. Moreover, to tackle the challenge of limited channel measurement in the THz band, the transfer learning technique is introduced in T-GAN, which reduces the size requirement of channel dataset for training and enhances the performance of channel modeling, through transferring the knowledge stored in a pre-trained model to a new model [12, 13]. Furthermore, the performance of T-GAN in modeling the channel distribution is validated by real measurements [14].

The contributions of this paper are listed as follows.

  • We propose a T-GAN based THz channel modeling method, in which a GAN is designed to capture the distribution of PDPs of the THz channel, by training on the dataset of PDP samples.

  • To tackle the challenge of limited measurement data for THz channel modeling, transfer learning is further exploited by T-GAN, which reduces the size requirement of training dataset, and enhances the performance of GAN, through transferring the knowledge stored in a pre-trained model to a new model.

The rest of the sections are organized as follows. Sec. II details the proposed T-GAN based channel modeling method. Sec. III demonstrates the performance of the proposed T-GAN method. The paper is concluded in Sec. IV.

Notation: aa is a scalar. a denotes a vector. A represents a matrix. 𝔼{}\mathbb{E}\{\cdot\} describes the expectation. \nabla denotes the gradient operation. \left\|\cdot\right\| represent the L2 norm. IN\textbf{I}_{N} defines an NN dimensional identity matrix. 𝒩\mathcal{N} denotes the normal distribution.

II Transfer GAN (T-GAN) Based Channel Modeling

In this section, the channel modeling problem is first formulated into a channel distribution learning problem. Then, the proposed GAN in T-GAN method is elaborated. Finally, T-GAN is presented.

II-A Problem Formulation

The channel impulse response (CIR) can be represented as

h(τ)=l=0L1αlejϕlδ(ττl),h(\tau)=\sum_{l=0}^{L-1}\alpha_{l}e^{j\phi_{l}}\delta(\tau-\tau_{l}), (1)

where τl\tau_{l} denotes the delay of the lthl^{th} multi-path components (MPCs), LL denotes the number of MPC, αl\alpha_{l} refers to the path gain and ϕl\phi_{l} represents the phase of the corresponding MPC. To characterize the channel, PDP is an important feature, which indicates the dispersion of power over the time delay, specifically, the received power with respect to the delay in a multi-path channel. It can be extracted from the channel impulse response by

P(τ)=|h(τ)|2,P(\tau)=|h(\tau)|^{2}, (2)

Then, the channel modeling problem is exploited by learning the distribution of PDPs denoted by prp_{r}, which is difficult to be analytically represented. Instead, the distribution prp_{r} can be captured by generating fake PDP samples with distribution pgp_{g}, such that the generated distribution pgp_{g} of PDPs can match the actual distribution prp_{r}.

II-B Proposed GAN

The GAN can be utilized to learn the distribution of PDPs denoted by prp_{r}, with the framework depicted in Fig 1. The GAN consists of two sub-networks, namely, generator and discriminator. The generator is aimed at generating fake samples G(z)G(z) to fool the discriminator, in which zz is the noise sample, by mapping the input noise distribution pz(z)p_{z}(z) to the generated distribution pg=p(G(z))p_{g}=p(G(z)). The discriminator tries to distinguish between real samples xx from prp_{r} and fake samples G(z)G(z) from pgp_{g}, and the output of the discriminator D(x)D(x) and D(G(z))D(G(z)) can be treated as the probability of being a real sample. The two networks are trained in an adversarial manner, which can be considered as a two-player zero-sum minimax game. Specifically, the training objective can be represented by

minGmaxD𝔼𝒙pr[logD(𝒙)]+𝔼𝒛pz[log(1D(G(𝒛)))],\mathop{\min}\limits_{G}\mathop{\max}\limits_{D}\mathbb{E}_{\boldsymbol{x}\sim p_{r}}[\log D(\boldsymbol{x})]+\mathbb{E}_{\boldsymbol{z}\sim p_{z}}[\log(1-D(G(\boldsymbol{z})))], (3)
Refer to caption
Figure 1: Framework of GAN.

where the generator minimizes the probability (1D(G(z))(1-D(G(z)) that the generated sample is detected as fake by the discriminator, while the discriminator maximizes this probability. Therefore, the generator and discriminator compete against each other with the opposite objectives in the training process. Through the adversarial training, the Nash equilibrium can be achieved, such that the generator and discriminator cannot improve their objectives by changing only their own network. Moreover, the global optimum of the training objective can be achieved in the equilibrium when pg=prp_{g}=p_{r}. However, training with the objective function in (3) is unstable, since the training objective is potentially not continuous with respect to the generator’s parameters [15]. Therefore, the improved version of GAN, namely, Wasserstein GAN with gradient penalty (WGAN-GP) [15] is adopted. The modified objective function is expressed as

minGmaxD𝔼𝒙pr[D(𝒙)]+\displaystyle\mathop{\min}\limits_{G}\mathop{\max}\limits_{D}\mathbb{E}_{\boldsymbol{x}\sim p_{r}}[D(\boldsymbol{x})]+ 𝔼𝒛pz[(1D(G(𝒛)))]\displaystyle\mathbb{E}_{\boldsymbol{z}\sim p_{z}}[(1-D(G(\boldsymbol{z})))] (4)
+\displaystyle+ λ𝔼𝒙~[(𝒙~D(𝒙~)1)2)],\displaystyle\lambda\mathbb{E}_{\tilde{\boldsymbol{x}}}[(\left\|\nabla_{\tilde{\boldsymbol{x}}}D(\tilde{\boldsymbol{x}})\right\|-1)^{2})],

where the last term is the gradient penalty term to enforce Lipschitz constraint that the gradient of the GAN network is upper-bounded by a maximum value, the symbol 𝒙~\tilde{\boldsymbol{x}} is the uniformly sampled point between the points of 𝒙\boldsymbol{x} and G(𝒛)G(\boldsymbol{z}). Moreover, the parameter λ\lambda is the penalty coefficient.

After introducing the framework of GAN, the detailed architecture of proposed GAN network is presented. The structures of generator G and discriminator D are depicted in Fig. 2, where the number in the bracket denotes the dimension. The input to the generator is a noise vector zz with dimension nz=100n_{z}=100, which is sampled from the probability density function 𝒩(0,σ2𝐈nz)\mathcal{N}(0,\sigma^{2}\mathbf{I}_{n_{z}}). The generator consists of five dense layers, and the numbers of neurons in the dense layers are 128, 128, 128, 128, 401, respectively. It is worth noting that the size of the output layer is equal to the size of PDP. The activation function of the first four dense layers is LeakyReLU function, which can speed up the convergence and avoid the gradient vanishing problem. The formula of the LeakyReLU function is expressed as

f(x)={x,ifx0αx,ifx<0,f(x)=\begin{cases}x,&\mathrm{if}\ x\geq 0\\ \alpha x,&\mathrm{if}\ x<0\end{cases}, (5)

where α\alpha is the slope coefficient when the value of neuron xx is negative. In addition to the LeakyReLU function, a Sigmoid function is utilized in the last layer, which maps the output to the range of [0, 1]. The Sigmoid function is defined as

f(x)=11+ex.f(x)=\frac{1}{1+e^{-x}}. (6)

After going through the dense layers and activation functions in the generator, the input noise vectors are transformed into the generated samples. Then, the generated samples together with real samples are passed to the discriminator.

The discriminator is designed to distinguish between generated samples and real samples. The numbers of neurons for the five dense layers in the discriminator are 512, 256, 128, 64, 1, respectively. The activation function chosen for the first 4 layers is the LeakyReLU function introduced before. The activation function for the output layer is linear activation function, which is decided by the objective function of WGAN-GP introduced in (4).

Refer to caption
Figure 2: Structure of generator and discriminator.
Refer to caption
Figure 3: Framework for T-GAN.

II-C Proposed T-GAN

The framework for the proposed T-GAN is depicted in Fig. 3, in which the transfer learning is conducted between the measurement and 3GPP TR 38.901 model [16]. The measured PDPs denote the PDPs extracted from measurement with a small size, while the simulated PDPs refer to the PDPs simulated by the 3GPP model, which is implemented with the extracted statistics from measurement. Then, the proposed GAN and T-GAN with the same network structure, are trained on the simulated PDPs and measured PDPs, respectively, to capture the distribution of PDPs. Since the size of measured PDPs is quite small for the training of T-GAN, which can cause the difficulty of converging or the over-fitting problem, the transfer learning is exploited to tackle these problems.

To describe the transfer learning formally, a domain denoted by 𝒟\mathcal{D} consists of a feature space 𝒳\mathcal{X} and a marginal probability distribution 𝒫(𝐗)\mathcal{P}(\mathbf{X}) defined on 𝐗={𝐱1,𝐱2,,𝐱N}𝒳\mathbf{X}=\{\mathbf{x}_{1},\mathbf{x}_{2},\cdots,\mathbf{x}_{N}\}\in\mathcal{X}, where NN is the number of feature vectors in 𝐗\mathbf{X}. As depicted in Fig. 3, the target domain 𝒟t\mathcal{D}_{t} and source domain 𝒟s\mathcal{D}_{s} are defined on measurement and 3GPP model, respectively. The feature spaces for the two domains are both constructed by PDPs, with different marginal probability distributions defined on measured PDPs 𝐗𝐭\mathbf{X_{t}} and simulated PDPs 𝐗𝐬\mathbf{X_{s}}.

Moreover, given a domain 𝒟(𝒳,P(𝐗))\mathcal{D}(\mathcal{X},P(\mathbf{X})), a task denoted by 𝒯\mathcal{T} is defined by a label space \mathcal{L} and a predictive function f()f(\cdot), and the predictive function is learned from the pairs (𝐱n,ln)(\mathbf{x}_{n},l_{n}) with 𝐱n𝐗\mathbf{x}_{n}\in\mathbf{X} and lnl_{n}\in\mathcal{L}. In the target domain 𝒟t\mathcal{D}_{t} and source domain 𝒟s\mathcal{D}_{s}, the tasks are the same to capture the distribution of PDPs, and the label space is ={0,1}\mathcal{L}=\{0,1\} representing whether the PDP sample is generated by the proposed GAN or from the training dataset. The T-GAN and GAN serve as the predictive functions ftf_{t} and fsf_{s}. Then, transfer learning is aimed at learning the function ftf_{t} in target domain 𝒟t\mathcal{D}_{t} with the knowledge of 𝒯s\mathcal{T}_{s} in source domain 𝒟s\mathcal{D}_{s}, i.e., transferring the knowledge stored in GAN trained on simulated PDPs to T-GAN trained on the measured PDPs.

The method of fine-tuning [13] is adopted for the transfer learning. The T-GAN is initialized with the weights of the GAN trained on the simulated PDPs, and is then fine-tuned on the measured PDPs with small size. It is worth noting that both the generator and discriminator in the GAN are transferred, which can yield the better performance in generating high quality samples and fast convergence, compared with transferring only the generator or the discriminator [13].

With transfer learning, the performance of T-GAN can be largely enhanced. Specifically, the channel statistics extracted for 3GPP method are captured by the proposed GAN trained on simulated PDPs, which are further transferred to T-GAN. Moreover, T-GAN can learn the features of PDPs that are not captured by 3GPP method, directly from measurement, which further improves the performance of T-GAN in modeling the distribution of PDPs.

III Experiment and Performance Evaluation

In this section, the experiment settings are elaborated. Moreover, the performance of the T-GAN are evaluated by comparing the generated distribution of PDPs with measurement.

III-A Dataset and Setup

The dataset is collected from the measurement campaign in [14]. which is conducted in an indoor corridor scenario at 306-321 GHz with 400 ns maximum delay, as depicted in Fig. 4. With the measurement data, the PDPs can be extracted to characterize the channel in the 21 receiver points. Since the sample frequency interval is relatively small, as 2.5 MHz, the measured PDPs are very long, including 6001 sample points, which results in extraordinary computation and time consumption to train the GANs. To address this problem, we only use the measured channel transfer functions in the frequency band from 314 to 315 GHz, based on which the PDPs can be shorten to 401 sample points.

The PDPs of the 21 measured channels make up the measured dataset. In addition to the measured dataset, the dataset of simulated PDPs can be generated by 3GPP model with the extracted statistics from the measurement, which consists of 10000 channels. Compared to the measured dataset, the simulated dataset has larger data size with the channel statistics embedded. Moreover, the PDPs in two datasets are normalized into the range of [0, 1] by the min-max normalization method.

The training procedure of the GAN network is explained in detail as follows. Firstly, the input noise vector zz of size 100 is generated by the multivariate normal distribution, which can provide the capabilities to transform into the desired distribution. The gradient penalty parameter λ\lambda in (4) is set as 10, which works well in the training process. Moreover, the stochastic gradient descent (SGD) optimizer is applied for the generator network, and the adaptive moment estimation (Adam) optimizer is chosen for the discriminator network. In addition, the learning rates of the two optimizers are both set as 0.0002 to stabilize the training.

All the experimental results are implemented on a PC with AMD Ryzen Threadripper 3990X @ 2.19 GHz and four Nvidia GeForce RTX 3090 Ti GPUs. In addition, the training of GAN network is carried out in the Pytorch framework.

Refer to caption
Figure 4: Measurement layout in the indoor corridor scenario [14].

III-B Convergence

The proposed GAN is first trained on the simulated dataset, and is then fine-tuned on the measured dataset with transfer learning to develop the T-GAN. The numbers of epochs for training the proposed GAN and T-GAN are both set as 10000. A epoch is defined as a complete training cycle through the training dataset, in which the generator and discriminator are iteratively trained for once. To demonstrate the benefits of transfer learning, the GAN is also trained on the measured dataset without transfer learning for comparison. The loss of generator denoted by G_loss and loss of discriminator denoted by D_loss are shown in the Fig. 5, in which the TG_loss and TD_loss correspond to the losses for T-GAN. For the simulated dataset, it is clear that the generator and discriminator reach the equilibrium in the end. For the measured dataset, the loss of T-GAN is close to the loss for the simulated dataset except for some small fluctuations. The fluctuations are due to the small size of the measured dataset. By comparison, the training is not stable for the GAN network without transfer leaning. There is large fluctuation in the discriminator loss, and the absolute values of G_loss and D_loss are quite large compared to the losses for the simulated dataset. The comparison demonstrates the benefits of the transfer learning in the training of GAN network, which enables T-GAN to converge with a small training dataset. Moreover, it takes only 4000 epochs for T-GAN to converge, compared to 6000 epochs for GAN trained on the simulated dataset. The training time of T-GAN on the measured dataset is also small, which is only 114 seconds compared to 7 hours for GAN trained on the simulated dataset. From these results, it is clear that the transfer learning technique can improve the convergence rate of T-GAN, and reduce the training overhead with the knowledge from the pre-trained model.

Refer to caption
Figure 5: Loss of the generator and discriminator in the GAN network.
Refer to caption
(a) Samples of PDP.
Refer to caption
(b) Average PDP.
Figure 6: Plot of PDPs generated by measurement, 3GPP, the proposed GAN and T-GAN.
Refer to caption
Figure 7: SSIM of PDP for 3GPP, the proposed GAN and T-GAN.
Refer to caption
Figure 8: Delay spread for 3GPP, the proposed GAN and T-GAN.

III-C Power Delay Profile

In the experiment, the samples of PDP from measurement, 3GPP method, the proposed GAN and T-GAN are compared as in Fig. 6(a). It is clear that the PDPs are similar to each other, which proves that the proposed GAN and T-GAN can learn the features of PDPs. Moreover, it is observed that PDP of measurement is more complex than PDP of 3GPP method. There are more peaks and fluctuations in the temporal domain. This shows that 3GPP cannot well capture the channel effects embedded in PDP. Comparing PDPs generated by the proposed GAN and T-GAN, the PDP generated by T-GAN is close to measurement, while the PDP generated by the proposed GAN is similar to the 3GPP approach. This is reasonable, since the T-GAN can capture the features of PDP from measurement through transfer learning, while the propose GAN can only learn the features of the simulated PDPs by 3GPP method.

In addition, the average PDPs for these method are plotted in Fig. 6(b). It is clear that T-GAN shows good agreement with measurement, while 3GPP and GAN have large deviations from measurement. The deviations can be measured by root-mean-square error (RMSE), calculated as

RMSE=1Nτ(Pm(iΔτ)Pg(iΔτ))2,\mathrm{RMSE}=\sqrt{\frac{1}{N_{\tau}}\sum(P_{m}(i\Delta\tau)-P_{g}(i\Delta\tau))^{2}}, (7)

where NτN_{\tau} denotes the number of sampling points in PDP, ii indexs temporal sample points of PDPs, NτN_{\tau} represents the number of sampling points and Δτ\Delta\tau is the sampling interval. Moreover, Pm(iΔτ)P_{m}(i\Delta\tau) and Pg(iΔτ)P_{g}(i\Delta\tau) are the average power in the ithi^{\text{th}} sample point of measured PDPs and generated PDPs, respectively. The results of RMSE for 3GPP, the proposed GAN and T-GAN are 4.29 dB, 4.12 dB and -4.82 dB, respectively. The T-GAN improves the performance of RMSE by about 9 dB, compared with other methods, which demonstrates that the T-GAN outperforms the other methods in terms of modeling the average power of PDP. This is attributed to the powerful capability of GAN in modeling the complex distribution, and the benefits of transfer learning in better utilizing the small measurement dataset.

Moreover, to measure the similarity quantitatively, Structure Similarity Index Measure (SSIM) is introduced, which is widely applied to evaluate the quality and similarity of images. The range of SSIM is from 0 to 1, and the value of SSIM is larger when the similarity between images is higher. The PDPs generated by 3GPP method, the proposed GAN and T-GAN are compared with measurement. The cumulative probability functions (CDFs) of SSIM for these method are shown in Fig. 7. It can be observed that the proposed T-GAN can achieve higher SSIM values compared with other methods. More than 40%\% of SSIM values are higher than 0.6 for T-GAN, compared to only 20%\% for 3GPP and the proposed GAN. This further demonstrates the better performance of T-GAN in modeling the PDPs.

III-D Delay Spread

Delay spread characterizes the power dispersion of multi-path components in the temporal domain, which can be calculated as the second central moment of PDPs, by

τ¯=i=0NτiΔτP(iΔτ)Δτi=0NτP(iΔτ)Δτ,τrms=i=0Nτ(iΔττ¯)2P(iΔτ)Δτi=0NτP(iΔτ)Δτ,\begin{split}\bar{\tau}&=\frac{\sum_{i=0}^{N_{\tau}}i\Delta\tau P(i\Delta\tau)\Delta\tau}{\sum_{i=0}^{N_{\tau}}P(i\Delta\tau)\Delta\tau},\\ \tau_{rms}&=\sqrt{\frac{\sum_{i=0}^{N_{\tau}}(i\Delta\tau-\bar{\tau})^{2}P(i\Delta\tau)\Delta\tau}{\sum_{i=0}^{N_{\tau}}P(i\Delta\tau)\Delta\tau}},\end{split} (8)

where τ¯\bar{\tau} denotes the mean delay weighted by the power, τrms\tau_{rms} refers to the root-mean-square (RMS) delay spread, and P(iΔτ)P(i\Delta\tau) are the power in the ithi^{\text{th}} sample point of PDPs.

Then, the CDF plot of delay spread for measurement, 3GPP, the proposed GAN and T-GAN is depicted in Fig. 8. It can be observed that the CDFs of delay spread for 3GPP, the proposed GAN and T-GAN match the measurement well.

IV Conclusion

In this paper, we proposed a T-GAN based THz channel modeling method, which can capture the distribution of PDPs for the THz channel with the designed GAN. Moreover, the transfer learning is exploited in T-GAN to reduce the size requirement of training dataset and enhance the performance of GAN, through transferring the knowledge stored in the pre-trained GAN on the simulated dataset to the target T-GAN trained on limited measurement. Finally, we validate the performance of T-GAN with measurement. T-GAN can generate the PDPs that have good agreement with measurement. Compared with conventional methods, T-GAN has better performance in modeling the distribution of PDPs, with 9 dB improved RMSE and higher SSIM. More than 40%\% of SSIM values are higher than 0.6 for T-GAN, compared to only 20% for 3GPP and the proposed GAN.

References

  • [1] I. F. Akyildiz, C. Han, Z. Hu, S. Nie, and J. M. Jornet, “Terahertz band communication: An old problem revisited and research directions for the next decade (invited paper),” IEEE Trans. Commun., vol. 70, no. 6, pp. 4250–4285, 2022.
  • [2] Z. Chen et al., “Terahertz wireless communications for 2030 and beyond: A cutting-edge frontier,” IEEE Commun. Mag., vol. 59, no. 11, pp. 66–72, 2021.
  • [3] C. Han, Y. Wu, Z. Chen, Y. Chen, and G. Wang, “THz ISAC: A Physical-Layer Perspective of Terahertz Integrated Sensing and Communication,” arXiv preprint:2209.03145, 2022.
  • [4] P. Petrus, J. Reed, and T. Rappaport, “Geometrical-based statistical macrocell channel model for mobile environments,” IEEE Trans. Commun., vol. 50, no. 3, pp. 495–502, 2002.
  • [5] Y. Chen, L. Yan, and C. Han, “Hybrid spherical- and planar-wave modeling and DCNN-powered estimation of Terahertz Ultra-Massive MIMO channels,” IEEE Trans. Commun., vol. 69, no. 10, pp. 7063–7076, 2021.
  • [6] Y. Chen, L. Yan, C. Han, and M. Tao, “Millidegree-level direction-of-arrival estimation and tracking for terahertz ultra-massive mimo systems,” IEEE Trans. Wirel. Commun., vol. 21, no. 2, pp. 869–883, 2022.
  • [7] T. Wang, C.-K. Wen, S. Jin, and G. Y. Li, “Deep learning-based CSI feedback approach for time-varying massive MIMO channels,” IEEE Wirel. Commun. Lett., vol. 8, no. 2, pp. 416–419, 2019.
  • [8] T. J. O’Shea, T. Roy, and N. West, “Approximating the void: Learning stochastic channel models from observation with variational generative adversarial networks,” in Proc. Int. Conf. Comput., Netw. Commun., 2019, pp. 681–686.
  • [9] H. Xiao, W. Tian, W. Liu, and J. Shen, “ChannelGan: Deep learning-based channel modeling and generating,” IEEE Wirel. Commun. Lett., vol. 11, no. 3, pp. 650–654, 2022.
  • [10] T. Orekondy, A. Behboodi, and J. B. Soriaga, “MIMO-GAN: Generative MIMO channel modeling,” in Proc. IEEE Int. Conf. Commun., 2022, pp. 5322–5328.
  • [11] Y. Wei, M.-M. Zhao, and M.-J. Zhao, “Channel distribution learning: Model-driven gan-based channel modeling for irs-aided wireless communication,” IEEE Trans. Commun., vol. 70, no. 7, pp. 4482–4497, 2022.
  • [12] N. V. Huynh and G. Y. Li, “Transfer learning for signal detection in wireless networks,” IEEE Wirel. Commun. Lett., pp. 1–1, 2022.
  • [13] Y. Wang, C. Wu, L. Herranz, J. van de Weijer, A. Gonzalez-Garcia, and B. Raducanu, “Transferring gans: generating images from limited data,” in ECCV, 2018.
  • [14] Y. Li, Y. Wang, Y. Chen, Z. Yu, and C. Han, “Channel measurement and analysis in an indoor corridor scenario at 300 ghz,” in Proc. IEEE Int. Conf. Commun., 2022, pp. 2888–2893.
  • [15] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. Courville, “Improved training of Wasserstein GANs,” in Proc. Int. Conf. Neural Inf. Process. Syst., 2017, p. 5769–5779.
  • [16] Study on Channel Model for Frequencies From 0.5 to 100 GHz (Release 15), document TR 38.901, 3GPP, 2018.