Chengfang Yue , Hui Tang , Jun Yang and Li ChaiA Generalized CNN Model with Automatic Hyperparameter Tuning for Millimeter Wave Channel PredictionAbstract: This paper focuses on millimeter wave (mmWave) channel prediction by machine learning (ML) methods. Previous ML-based mmWave channel predictors have limitations on requirements of the amount of training data, model generalization ability, robustness to noise, etc. In this paper, we propose a CNN model with a novel feature selection strategy for mmWave channel prediction. Automatic hyperparameter tuning (AHT) algorithms are embedded in the training process to iteratively optimize the predictive performance of the proposed CNN. The diversification strategy is leveraged to enhance the robustness of the AHT procedure against different communication scenarios. To improve the generalization ability of the prediction model, the input features are designed to capture the correlation between the physical environment and the channel characteristics. In parallel, the Cartesian coordinates of the transmitter (Tx) and receiver (Rx) are transformed into polar ones to reduce the model’s sensitivity to coordinate noise. Numerical results demonstrate the effectiveness of the proposed CNN model in predicting mmWave channel characteristics in various communication scenarios. Keywords: Automatic hyperparameter tuning , channel prediction , CNN , diversification strategy , feature selection , mmWave communication I. INTRODUCTIONTHE design of wireless communication systems in specific scenarios relies heavily on knowledge of the channels [1]. However, in millimeter wave (mmWave) communication systems, the implementation of hybrid analog/digital array architecture makes it challenging to obtain the channels directly [2]. Traditional channel modeling methodologies, such as deterministic or stochastic, face significant challenges when applied to mmWave channels, including high implementation overheads, heavy computational burden, and increased data traffic [3]. Alternatively, channels can be studied from the perspective of parameters since channel parameters aid in the design of mmWave communication systems. For instance, path loss helps understand the coverage of a base station, time delay indicates communication latency, and angles of arrival and departure contribute to forming highly directional beams. Machine learning (ML) is a powerful tool for channel modeling and parameter prediction due to its self-learning, non-linear fitting, and big data processing capabilities [4]–[11]. For example, [5] proposed an artificial neural network (ANN) based model to playback measured channels, where threedimensional coordinates of antennas and carrier frequency were selected as input features, while cross-polarization ratio (XPR), amplitude, delay, and phase were chosen as output labels. [7]–[9] applied neural network (NN) [7], relevance vector machine (RVM) [8], and multi-layer perceptron (MLP) [9] to predict the large-scale path loss (PL) and received power (RP), respectively. In [7]–[9], the height of Tx and Rx antennas, the distance of Tx and Rx antennas, carrier frequency, and diffraction loss were selected as input features to learn and predict PL and RP. The RVM model was utilized to estimate the angle of arrival (AoA) [10]. The procedure was to estimate the rough positions of the signals by sparse RVM and then to obtain the precise estimated angles by searching the preset spatial grid. In [11], CNN was used to predict mmWave channel parameters using coordinates of transmitters (Txs) and receivers (Rxs), highlighting the effect of data acquisition methods on prediction performance. The generalization ability of most state-of-the-art channel prediction models is inadequate, which hampers the availability of ML-based channel prediction methods. Specifically, deploying a trained channel prediction model in new communication systems necessitates conducting repeated channel soundings to extract channel parameters and build new training databases. However, channel sounding is both costly and challenging to implement, particularly for mmWave communication systems. To reduce the labeling overhead associated with channel sounding and facilitate ML-based channel prediction, one approach is to devise a model that is adaptable to diverse communication scenarios. On the other hand, it is well-known that the empirical performance of ML models relies significantly on hyperparameters, which should be carefully tuned. However, manual hyperparameter tuning is a challenging and tedious task that often leads to suboptimal solutions [12]. To address this issue, various automatic hyperparameter optimization (AHT) techniques were proposed, such as random search, Bayesian optimization, and Gaussian process [12]–[17]. In the field of ML-based channel prediction, the robustness of the AHT process over different datasets is relatively under-researched. Coordinates of Txs and Rxs were commonly selected as input features for channel modeling and prediction [5], [11], [18]. In situations where the positions of antennas fluctuate dynamically (e.g., antennas mounted on drones and unmanned surface vehicles), coordinate noise can not be ignored. However, these problems have not been well addressed in existing literature. This paper introduces an automatic hyperparameter tuning algorithm-assisted CNN model (AHT-assisted CNN) to predict various channel parameters. Our work aims to address the challenges of reducing offline training efforts, improving the applicability of the prediction model to different communication systems with less channel sounding overhead, and enhancing the robustness of the framework to coordinate noise. The main contributions of this study can be summarized as follows: · The AHT-assisted CNN model proposes the integration of AHT algorithms into the CNN training process to optimize the hyperparameters in an iterative manner by considering the tuning process as an optimization problem of the unknown black-box function [13], [14]. The AHT procedure is limited by inherent vulnerability, which means a single tuning algorithm might converge rapidly for one database but then perform poorly for another [17]. To overcome this limitation, we take advantage of the diversification strategy to make the AHT procedure robust across different databases, thus saving offline training effort. · We save the labeling overhead associated with channel sounding by improving the generalization ability of the CNN model and reducing the amount of training data required. To achieve this purpose, the input features are designed according to the mmWave propagation mechanism to capture the intrinsic interaction between the physical environment and the corresponding channel. Consequently, the input features are determined as the received power, azimuth mean angle of departure, elevation mean angle of departure, and coordinates of Txs and Rxs, which allows us to predict the path loss, azimuth mean angle of arrival, elevation mean angle of arrival, and delay mean. Numerical results show that such a feature selection strategy improves the generalization ability of the channel prediction model and requires less training data to achieve satisfactory prediction accuracy. · The recorded Cartesian coordinates of the Txs and Rxs antennas are transformed into polar ones to enable the model to be robust to coordinate noise. We demonstrate that the featured polar coordinates are less sensitive to coordinate noise, which improves the predictive performance of the model in situations where the actual coordinates of the antennas fluctuate. To create simulation datasets, the ray tracing software Wireless InSite [19] is utilized. Ray tracing is a well-established deterministic technique that is commonly used for modeling radio propagation. This technique is based on the principles of geometrical optics (GO) and the uniform theory of diffraction (UTD). Various interactions between rays and objects, such as reflection, transmission, scattering, and diffraction, can be categorized using this approach. Numerical results show that the presented AHT-assisted CNN can achieve state-ofthe- art performance in various aspects, including less training efforts, smaller training datasets, stronger generalization ability, and robustness to coordinate noise. We present the abbreviations (ABB) and corresponding full names (FN) of the involved channel parameters in Table I to avoid confusion. II. SYSTEM MODELWe consider an mmWave communication system with K randomly distributed Rxs and the same number of Txs, where both Txs and Rxs are equipped with a single antenna. The operating frequency is 100 GHz. M Tx-Rx pairs are established, where M equals K × K. Then the received signal of the mth Tx-Rx pair can be written as
where x denotes the transmitted signal, [TeX:] $$n_m \sim \mathcal{C N}\left(\mu, \sigma^2\right)$$ is the additive white Gaussian noise at user [TeX:] $$m, \sigma^2$$ denotes the variance of noise, [TeX:] $$h_m(f)$$ represent the channel for the mth Tx-Rx pair. Let f represent the carrier frequency, [TeX:] $$\alpha_{m l}, \tau_{m l}, \beta_{m l}, \phi_{m l}, \text { and } \theta_{m l}$$ refer to the attenuation, time delay, phase shift, azimuth AoAs, and elevation AoAs associated with the lth path on the mth channel, respectively. Let L be the number of paths, and [TeX:] $$a_m\left(\phi_{m l}, \theta_{m l}\right)$$ be the mth antenna response vector. As in [20], the channel for the mth Tx-Rx pair is defined as
(2)[TeX:] $$h_m(f)=\sum_{l=1}^L \alpha_{m l} e^{-j 2 \pi f \tau_{m l}+j \beta_{m l}} \boldsymbol{a}_m\left(\phi_{m l}, \theta_{m l}\right),$$where [TeX:] $$m=1,2, \cdots, M .$$ By denoting the path loss (PL), azimuth mean angle of arrival (AMAoA), elevation mean angle of arrival (EMAoA), and delay mean (DM) of the mth channel by [TeX:] $$\rho_m, \bar{\phi}_m, \bar{\theta}_m, \text { and } \bar{\tau}_m,$$ respectively, we define the mapping from [TeX:] $$h_m(f)$$ to channel characteristics according to (2) as follows,
(3)[TeX:] $$h_m(f)=\sum_{l=1}^L \alpha_{m l} e^{-j 2 \pi f \tau_{m l}+j \beta_{m l}} \boldsymbol{a}_m\left(\phi_{m l}, \theta_{m l}\right),$$It is worth noting that we choose the mean arrival angles and mean time delay to define the above mapping function because the time delay and angles of each path are difficult to obtain and process during the model training procedure.Alternatively, statistical channel characteristics such as angle spread and delay spread can be used to model channels instead of mean angles and mean delay. Similarly, the mean values of the departure angles (AMAoD and EMAoD) are used in this paper. The phase shift [TeX:] $$\beta_i$$ is not considered in (3) because multipath phases of the channel are immeasurable inpractice and are assumed to be uniformly distributed in the range [TeX:] $$[0,2 \pi)$$ [21]. The channel [TeX:] $$h_m(f)$$ in a certain communication scenario is affected by the distribution of multipaths, which is determined by the environment and antenna coordinates. As illustrated in Fig. 1, different multipaths are distinguished by AoDs, AoAs, and RP. In view of this, we choose coordinates, AoDs, AoAs, and RP as environment factors and define another mapping between the environment factors and the corresponding channel as
(4)[TeX:] $$\mathcal{F}_*:\left[C_{\boldsymbol{m}}, \bar{\varphi}_m, \bar{\vartheta}_m, p_m\right] \rightarrow h_m(f),$$where [TeX:] $$C_m$$ denotes the vector of Cartesian coordinates of the mth Tx-Rx pair, that is [TeX:] $$C_m=\left[{x_m}^T, {y_m}^T, {x_m}^R, {y_m}^R\right],$$ while [TeX:] $$\bar{\varphi}_m, \bar{\vartheta}_m, \text { and } p_m$$ represent AMAoD, EMAoD and RP of the mth Tx-Rx channel, respectively. By combining (3) and (4), we obtain the mapping between the environment factors and the channel parameters as
(5)[TeX:] $$\tilde{\mathcal{F}}:\left[C_{\boldsymbol{m}}, \bar{\varphi}_m, \bar{\vartheta}_m, p_m\right] \rightarrow\left[\rho_m, \bar{\phi}_m, \bar{\theta}_m, \bar{\tau}_m\right].$$As the mapping [TeX:] $$\tilde{\mathcal{F}}$$ indicated, the channel parameters vector [TeX:] $$\left[\rho_m, \bar{\phi}_m, \bar{\theta}_m, \bar{\tau}_m\right]$$ is closely related to the environment factors [TeX:] $$\left[C_m, \bar{\varphi}_m, \bar{\vartheta}_m, p_m\right].$$ When designing practical communication systems in different scenarios, important properties such as base station coverage, antenna orientation, and communication latency, can be investigated with the knowledge of path loss, arrival and departure angles, and time delay instead of repeatedly estimating or sounding channels. Our goal is to obtain such channel parameters from a new perspective in situations where channels are difficult to estimate or repeatedly sound. Specifically, we aim to construct a generalized CNN model which can predict channel characteristics for various different communication scenarios by learning the mapping [TeX:] $$\tilde{\mathcal{F}}$$ from datasets. Furthermore, the model is robust to coordinate noise. III. CNN-BASED CHANNEL PREDICTIONThe mapping [TeX:] $$\tilde{\mathcal{F}}$$ defined in Section II represents the interaction between the communication scenario and the corresponding channel parameters. [TeX:] $$\tilde{\mathcal{F}}$$ includes various aspects of the communication process, such as the scenario geometry, materials, reflection, etc. Therefore, it is difficult to find an analytical model of [TeX:] $$\tilde{\mathcal{F}}$$. In this section, we propose a CNN model for learning the mapping [TeX:] $$\tilde{\mathcal{F}}$$. We first introduce the procedure of the feature selection and preprocessing, and how it contributes to improving the model’s generalization ability and robustness to coordinate noise. Afterwards, we explain the motivation and architecture of the CNN. A. Feature Selection and PreprocessingThe correlation between features and labels is considered to be highly relevant to the predictive performance of the model. One focus of this paper is to construct highly correlated input features and output labels according to the mmWave communication mechanism. In particular, it is well known that path loss is related to received power. The antenna response between the mth Tx-Rx pair [TeX:] $$\mathbf{A}\left(\Psi_m, \Theta_m\right)$$ is defined as follows,
(6)[TeX:] $$\mathbf{A}\left(\Psi_m, \Theta_m\right)=a_{\boldsymbol{R}}\left(\phi_m, \theta_m\right) {a_{\boldsymbol{T}}}^*\left(\varphi_m, \vartheta_m\right),$$where [TeX:] $$\phi_m \text { and } \theta_m$$ are the azimuth and elevation AoAs of the mth Tx-Rx pair, respectively, while [TeX:] $$\varphi_m \text { and } \vartheta_m$$ refer to the azimuth and elevation AoDs of the mth Tx-Rx pair, respectively [22]. Based on the fact that the AoD and AoA enjoy reciprocity [23], an intuitive inference from (6) is that the AoAs ([TeX:] $$\phi_m \text { and } \theta_m$$) are coupled with AoDs ([TeX:] $$\varphi_m \text { and } \vartheta_m$$). In addition, the coordinates indicate the positional relationship between the Tx and the corresponding Rx, which affects the distribution of channels. In summary, the environment factors [TeX:] $$[\boldsymbol{C}, \bar{\varphi}, \bar{\vartheta}, p]$$ are coupled with the channel characteristic vector [TeX:] $$[\rho, \bar{\phi}, \bar{\theta}, \bar{\tau}].$$ Hence, different from the existing literature, we choose [TeX:] $$[\boldsymbol{C}, \bar{\varphi}, \bar{\vartheta}, p]$$ and [TeX:] $$[\rho, \bar{\phi}, \bar{\theta}, \bar{\tau}].$$ as the input features and output labels, respectively, to enable the model to exploit the correlation between the featured and labeled channel parameters. The communication scenario (geometry, scattering, communication settings, etc.) determines the featured environment factors. Thus the ML model can learn the mapping between the environment information and the corresponding labeled channel parameters, allowing the model to adapt to different scenarios. It is necessary to clarify the reasonableness of the proposed feature selection strategy as follows. Although the path loss can be directly obtained via power calculation, we still predict the path loss using the received power for the following considerations. As depicted in Fig. 1, the received power is a metric of both the propagation distance and the scattering distribution. Specifically, multipath L1, L2, and L3 experience different propagation distances and scatters, where L1 is a reflective path, L2 is a non-LoS path, and L3 is an LoS path. The received powers resulting from the three paths (RP1, RP2, and RP3) are different, which means that we can distinguish one multipath from the other based on the received power. In other words, the received power is an indicator of the distribution of multipath. In addition, the received power is relatively easy to measure directly in practical communication systems. Fig. 8 shows that the featured received power contributes to not only accurately predicting the path loss but also improving the adaptivity of predicting the AoAs and time delay. Thus, RP is a reasonable featured environment factor. By the same token, elevation and azimuth angles of departure are chosen as environment factors. In this way, the proposed model can predict AoAs from known AoDs instead of jointly extracting AoAs and AoDs from the channel impulse response. For practical applications, the proposed model can be used to predict AoAs by certain predefined AoDs for mmWave communication systems equipped with directional antennas, which helps to arrange or adjust the direction of the receiving antennas in real time. Inspired by the wide application of polar coordinates in antenna theory, we transform the Cartesian coordinates of Txs and Rxs [TeX:] $$\boldsymbol{C}_{\boldsymbol{m}}=\left[{x_m}^T, {y_m}^T, {x_m}^R, {y_m}^R\right]$$ to the polar coordinates [TeX:] $$\boldsymbol{P}_{\boldsymbol{m}}=\left[r_m{ }^T, r_m{ }^R, \delta_m{}^T, \delta_m{}^R\right]$$ by the following equations,
(7)[TeX:] $$\left\{\begin{array}{l} r_m^T=\sqrt{\left(x_m^T\right)^2+\left(y_m^T\right)^2} \\ r_n^R=\sqrt{\left(x_m^R\right)^2+\left(y_m^R\right)^2} \\ \delta_m^T=\arctan \frac{y_m^T}{x_m^T} \\ \delta_m^R=\arctan \frac{y_m^R}{x_m^R} \end{array} .\right.$$It is widely known that PL and DM depend on the distance d between Tx and Rx, which is given by
(8)[TeX:] $$d=\left[\boldsymbol{C}_{\boldsymbol{m}} \boldsymbol{C}_{\boldsymbol{m}}^{\boldsymbol{T}}-2\left(x_m^T x_m^R-y_m^T y_m^R\right)\right]^{\frac{1}{2}},$$or
(9)[TeX:] $$d=\left[\left(r_m^T\right)^2+\left(r_m^R\right)^2-2 r_m^T r_m^R \cos \left(\delta_m^T-\delta_m^R\right)\right]^{\frac{1}{2}}.$$In (8), the distance is calculated by Cartesian coordinates. When [TeX:] $$x_m^T x_n^R-y_m^T y_n^R=0$$, d relates to 4 variables. The distance d calculated by polar coordinates in (9) depends on only 2 variables under the condition that [TeX:] $$\delta_m^T-\delta_n^R=0$$. In other words, the labeled channel parameters are determined by fewer featured coordinate variables in situations where the mobile antennas move to abnormal positions, making the model less sensitive to coordinate noise. B. The Motivation and Architecture of CNNWe propose a CNN rather than a regular fully-connected neural network (FCN) model for the following reasons. First, as discussed in Subsection III-A, the featured environment factors are coupled with the labeled channel parameters. With this in mind, the convolutional layers are leveraged to extract the correlation. In parallel, considering the characteristic of the feature dimension, CNN is a straightforward choice because of the ability to compress redundant channel data [11]. Specifically, the dimension of the training data is M × N, where M is the total number of Tx-Rx pairs, which is 90000 for the park and square scenarios and is 60000 for the lab scenario as detailed in Subsection IV-B. N denotes the number of input features. The schematic layout of the proposed model is presented in Fig. 2. The first convolutional layer filters the 1 × N input vector with [TeX:] $$\omega_2$$ kernels. The output of the first convolutional layer is fed to the second convolutional layer and filtered with [TeX:] $$\omega_3$$ kernels. The third convolutional layer processes the output of the second convolutional layer with [TeX:] $$\omega_4$$ kernels. The kernel size is 1 × S. The output of the third convolutional layer is then connected to three dense layers with [TeX:] $$\omega_5, \omega_6, \omega_7$$ neurons, respectively. To speed up the convergence rate, the learning rate [TeX:] $$\omega_1$$, the activation function F, and the batch size [TeX:] $$N_{b s}$$ are adjusted during the training process. IV. AHT-ASSISTED TRAINING PROCESSIn this section, we first highlight the procedure of AHT, followed by introducing the generation of training databases. Then, we describe in detail the training process of the proposed CNN. A. AHT Algorithms1) Formulation of AHT: As shown in Fig. 2, AHT algorithms are incorporated into the CNN training process to adjust hyperparameters to improve the training efficiency and obtain optimal solutions. To formulate the AHT-assisted training process, we first define a function [TeX:] $$\Delta$$ to represent the mapping between the hyperparameters and the validation loss. This function can be written as
where [TeX:] $$\boldsymbol{\Omega}=\left[\omega_1, \omega_2, \cdots, \omega_7, F, N_{b s}, S\right]$$ is the hyperparameter vector. The procedure of finding optimal hyperparameters [TeX:] $$\Omega_{\mathrm{opt}}$$ can be viewed as an optimization problem defined as
(11)[TeX:] $$\boldsymbol{\Omega}_{\mathrm{opt}} \triangleq \arg \min _{\boldsymbol{\Omega} \in \chi} \Delta(\boldsymbol{\Omega}),$$where [TeX:] $$\chi$$ denotes the search space of hyperparameters. The search space [TeX:] $$\chi$$ in this paper is investigated according to the experience of hyperparameter tuning and is given in Table II. It is worth noting that the search range can be arbitrarily modified for different application scenarios. The optimization problem in (11) is non-convex [17] and is solved by AHT algorithms instead of the tedious hand-tuning method. TABLE II HYPERPARAMETERS AND CORRESPONDING SEARCHING RANGE.
2) Procedure of AHT: There are two components to the AHT-assisted training algorithm. The first component is the probabilistic surrogate model, which includes a prior distribution and an observation model. The prior distribution describes our beliefs about the behavior of the target black-box function. The observation model captures the interaction between the hyperparameters and resulting validation loss [14]. The second one is called the acquisition function, which evaluates the candidate hyperparameters for the next observation. The AHT procedure is integrated with the diversification strategy to overcome the vulnerability of AHT algorithms to different databases. One of the popular diversification strategies is called ensemble, which enploys a set of learned models rather than apply a single model [24]. The proposed AHT procedure adopts diversification at model level as in the ensemble to produce optimal convergence under different scenarios. Specifically, we combine three surrogate models into the AHT algorithms, i.e., Gaussian processes (GP), gradientboosted regression trees (GBRT), and random forests (RF), to create three AHT algorithms, including GP-based Bayesian optimization (BO-GP), GBRT-based sequential model-based optimization (SMBO-GBRT) and RF based SMBO (SMBO-RF). Considering the trade-off between the training time and performance, we employ three different models in this paper. However, additional surrogate models can be conveniently added to our AHT algorithms for more complicated situations. Random search (RS) is employed as a baseline instead of grid search because the latter is very timeconsuming. The acquisition function is based on the expectation improvement (EI) strategy. Let [TeX:] $$\mu_n(\Omega), \sigma_n(\Omega), \Phi_C(\cdot), \text { and } \Phi_P(\cdot)$$ be the posterior mean function, posterior variance function, cumulative distribution function (CDF) and probability density function (PDF), respectively. Denote [TeX:] $$\tilde{\Delta}_n^*$$ as the best observation. The acquisition function is written as
(12)[TeX:] $$\tilde{\Delta}_n(\Omega)=\left\{\begin{array}{cc} \left(\mu_n(\Omega)-\tilde{\Delta}_n^*\right) \Phi_C(\Gamma) & \\ +\sigma_n(\Omega) \Phi_P(\Gamma), & \sigma_n(\Omega)\gt 0 \\ 0, & \sigma_n(\Omega)=0 \end{array},\right.$$where
a) Pseudocode for AHT: Pseudocode for the AHTassisted training algorithms is shown in Algorithm 1. During the training process, surrogate models are adopted in turn thanks to the sequential nature of the proposed AHT algorithms. The prior belief and the current observation [TeX:] $$\mathcal{H}_t$$ are prescribed by the probabilistic surrogate model. The observation set [TeX:] $$\mathcal{H}_t$$ consists of hyperparameters and the corresponding validation loss. For each surrogate model, the AHT algorithms incorporate prior belief to estimate [TeX:] $$\tilde{\Delta}(\boldsymbol{\Omega}).$$ The prior distribution is updated by observing the performance of the model under the current hyperparameters to obtain the posterior distribution. The posterior distribution is then substituted into the EI-based acquisition function in (12) to determine the next hyperparameters [TeX:] $$\Omega_{n+1}.$$ Then evaluate [TeX:] $$\Delta\left(\Omega_{n+1}\right)$$ with the real black-box function [TeX:] $$\Delta$$, followed by updating the set of observations [TeX:] $$\mathcal{H}_t$$. The probabilistic surrogate model is then updated with the posterior distribution and the set of observations and iterates sequentially until the early stopping criterion is met. In this paper, the stopping criterion is to end the iteration when no performance improvement is achieved for 10 consecutive calls. The observations [TeX:] $$\mathcal{H}_t$$ obtained by all T models are recorded in [TeX:] $$\mathcal{H}$$. The validation loss in each [TeX:] $$\mathcal{H}_t$$ is compared, and the final output of the AHT algorithm is the hyperparameter combination corresponding to the minimum validation loss. B. Training Data GenerationTo generate the training data, we build an outdoor scenario (a virtual park) and an indoor scenario (a virtual lab) in the accurate ray-tracing simulator. The layouts of the park and the lab are shown in Fig. 3. The park is [TeX:] $$400 \times 400 {~m}^2$$ and consists of a lake, a fountain, trees, grass, pavements, and constructions. The lab is furnished with tables, cabinets, and office furniture (bookshelf, electronic devices), and occupies an area of [TeX:] $$30 \times 30 {~m}^2$$. In the outdoor scenarios (the park and the square), the constructions are made of concrete, and the pavement is constructed using bricks. For the indoor scenario (the lab), wooden tables, chairs, and cabinets are used, while the ceiling, floor, and walls are made of concrete. The dielectric constants of the materials used in the study, including concrete, brick, wood, foliage, glass, office furniture (bookshelves and electronic devices), and water, are given in Table III [25]–[27]. The simulation frequency and bandwidth are set to 100 GHz and 2.5 GHz, respectively. A maximum of two orders of reflection and one order of diffraction are simulated due to the significant attenuation of mmWave propagation. Scattering due to surface roughness is neglected for the sake of the simplicity of simulation. The channel characteristics involved are calculated by
(14)[TeX:] $$\left\{\begin{array}{l} \bar{\tau}_m=\frac{\sum_{l=1}^L \tau_{m, l}}{L} \\ \bar{\varphi}_m=\frac{\sum_{l=1}^L \varphi_{m, l}}{L} \\ \bar{\vartheta}_m=\frac{\sum_{l=1}^L \vartheta_{m, l}}{L}, \\ \bar{\phi}_m=\frac{\sum_{l=1}^L \phi_{m, l}}{L} \\ \bar{\theta}_m=\frac{\sum_{l=1}^L \theta_{m, l}}{L} \end{array}\right.$$where l is the index of the propagation path and L denotes the total number of paths. We set the reference point at the lower left corner of each simulated communication area to obtain the 2-D relative coordinates of both Txs and Rxs. The coordinates are assumed to be calculated by highly accurate positioning techniques (e.g., via [28]). TABLE III DIELECTRIC CONSTANTS OF DIFFERENT MATERIALS AT 100 GHZ.
We randomly locate 300 Txs and 300 Rxs within the park to create 90000 Tx-Rx pairs and collect 90000 input features [TeX:] $$\left[\boldsymbol{C}_{\boldsymbol{m}}, \bar{\varphi}_m, \bar{\vartheta}_m, p_m\right]$$ and output labels [TeX:] $$\left[\rho_m, \bar{\phi}_m, \bar{\theta}_m, \bar{\tau}_m\right].$$ Due to the inherent sparsity of mmWave propagation, a maximum of 5 paths are considered. Each Tx (Rx) is equipped with a single half-wave dual-polarization antenna. The scenarios and the resulting databases are described in Table IV. The original scenario is the virtual park, where database i is collected, and then database ii is generated by transforming the Cartesian coordinates in database i to polar coordinates. We add coordinates noise equally to database i and database ii to generate database iii and database iv, respectively, while keeping the actual antenna positions unchanged to validate the robustness of the model to coordinate noise. A virtual square is then built as a similar environment for the park to generate database v. The only difference between the park and the square is the distribution of scatterers. The virtual indoor environment (the lab) is constructed as a different scenario from the park to obtain database vi. 300 Txs and 300 Rxs are randomly placed in the lab scenario to create 90000 Tx-Rx pairs. 60000 sets of valid featured and labeled channel data are collected. Both database v and database vi are built to validate the generalization ability of the model. TABLE IV DATABASES AND CORRESPONDING SCENARIOS.
When features and labels are highly correlated, the model can more easily learn the correlation between them, resulting in better prediction performance with a smaller sample of data [29]. To verify that the proposed feature selection strategy can reduce the amount of training data, the percentage of the training set varies from 10 percent to 70 percent at intervals of 10 during the training stage. The test and validation proportions are set equal. C. Training ProcessThe Adam optimizer is applied to the training of the network with a learning rate of [TeX:] $$\omega_1$$. The maximum iteration epoch is set to 3000, and early stopping with patience 100 is adopted to avoid overfitting. Also, the learning rate [TeX:] $$\omega_1$$ decays by a factor of 0.5 whenever the validation loss no longer diminishes for a consecutive 100 epochs. The CNN is trained to minimize the loss function formulated as
(15)[TeX:] $${ Loss }=\frac{1}{4 N_{B S}} \sum_{i=1}^4 \sum_{j=1}^{N_{B S}}\left(\hat{p}_{i j}-p_{i j}\right)^2 \text {, }$$where [TeX:] $$N_{b s}$$ is the batch size, [TeX:] $$\hat{p}_{i j} \text { and } p_{i j}$$ are the jth predicted and true values of the ith channel parameter (one of [TeX:] $$\rho, \bar{\phi}, \bar{\theta}, \text { and } \bar{\tau}$$), respectively. During the network training process, the hyperparameters listed in Table II should be tuned repeatedly until the validation loss is satisfactory. V. NUMERICAL RESULTSWe train three models with AHT algorithms to make comparisons, i.e., the enhanced CNN, the standard CNN, and a sixhidden- layer FCN without convolutional layers. The standard CNN is fed with Cartesian coordinates of Tx and Rx, which is the same as the model proposed in [11], while the enhanced one is the model presented in this paper, which is enhanced by the coordinate transformation and feature selection strategy. The input features and output labels of the FCN are the same as those of the enhanced CNN. We utilize a computer with an Intel Core i7-9700k central processing unit and an NVIDIA RTX 2070 graphics processing unit to train the network.The training time for each surrogate model amount to approximately 2 hours, while the average training time for a full model consisting of 3 surrogate models is less than 6 hours. A. AHT Procedure AnalysisWe train the enhanced model by database ii and database vi respectively, with the assistance of AHT algorithms and plot the convergence traces of: · GP as the surrogate model for BO (BO-GP), · GBRT as the surrogate model for SMBO (SMBO-GBRT), · RF as the surrogate model for SMBO (SMBO-RF), and · RS. The plots given as Figs. 4 and 5 show the value of the determination coefficient [TeX:] $$R^2$$ (y-axis) between the predicted and true channel parameters as a function of the number of iterations (x-axis). [TeX:] $$R^2$$ is an indicator of the prediction accuracy calculated by
(16)[TeX:] $$R^2=1-\frac{\sum_i\left(\hat{p}_{i j}-p_{i j}\right)^2}{\sum_i\left(p_{i j}-\bar{p}_i\right)^2} .$$[TeX:] $$R^2$$ varies in the range of [0,1]. In (16), [TeX:] $$\hat{p}_{i j} \text { and } p_{i j}$$ are the jth predicted and true values of the ith channel parameter (one of [TeX:] $$\rho, \bar{\phi}, \bar{\theta}, \text { and } \bar{\tau}$$), respectively, and [TeX:] $$\bar{p}_i$$ denotes the average value of the ith output channel parameter, calculated by [TeX:] $$\bar{p}_i=\sum_{v=1}^V \bar{p}_{i v} / V$$, with V denoting the total number of the ith labeled channel parameters [TeX:] $$p_i$$ in the validation set. The resulting [TeX:] $$R^2$$ scores of all predicted channel parameters are averaged with uniform weight to obtain the final [TeX:] $$R^2$$. The convergence rates and optimization performances of each tuning algorithm used vary for different databases, as shown in Figs. 4 and 5. This validates the necessity of the diversification strategy, which helps to determine the optimal AHT algorithm for different communication scenarios. Furthermore, due to its fast convergence rate and satisfactory accuracy, the proposed AHT procedure is suitable for embedding in the training process of the channel prediction CNN. This contributes greatly to saving training effort and improving prediction performance. Fig. 4. Convergence plot of AHT algorithms for the indoor scenario. (One iteration is defined as one call.) ![]() Fig. 5. Convergence plot of AHT algorithms for the outdoor scenario. (One iteration is defined as one call.) ![]() Table V presents the optimal hyperparameters obtained by the AHT algorithms for the park scenario. The resulting CNN model is used to plot the fitting curves between the predicted and real channel parameters, as shown in Fig. 6. From this plot, it is evident that the predicted values fit the real ones well. TABLE V OPTIMAL HYPERPARAMETERS OBTAINED BY AHT ALGORITHMS FOR THE PARK SCENARIO.
Fig. 6. Fitting plots for predicted channel parameters. The x-axis represents the number of samples. ![]() B. Impact of the Size of the Training Set and Robustness AnalysisThe standard CNN model is trained by the database i, whereas both the enhanced CNN and FCN are trained using the database ii. Afterward, the trained models are tested on database iii (the standard CNN) and database iv (the enhanced CNN and FCN), respectively, to validate the robustness of the model against coordinate noise. The figure presented in Fig. 7 illustrates the relationship between the coefficient of determination and the size of the training data. The results indicate that the enhanced AHT-assisted CNN is capable of achieving a decent [TeX:] $$R^2$$ value (higher than 0.8) with less training data compared to the standard CNN. This characteristic of the enhanced model makes it more robust to data size, which is highly advantageous for practical applications where data collection can be challenging. Additionally, both the enhanced model and the FCN exhibit significantly higher accuracy. This improved performance is attributed to the feature selection and preprocessing procedure employed in the enhanced model. Furthermore, the enhanced CNN exhibits robustness to coordinate noise, while the standard CNN is completely unreliable under such conditions. This robustness allows the proposed enhanced model to detect changes in communication scenarios when coordinate fluctuations occur. C. Generalization Ability of the Proposed ModelIn this subsection, we eveluate the generalization ability of the enhanced and standard models by testing the trained models on database v (the square scenario) and database vi (the lab scenario), respectively. Numerical results are given by Fig. 8. The enhanced CNN shows adaptability to communication scenarios, indicating that the proposed model can accurately predict channel parameters when the scenario dynamically changes to a similar one. Consequently,our proposed model can be deployed in similar communication scenarios without the need for re-training. As for the indoor scenario, all models require offline retraining. Nevertheless, the enhanced AHT-assisted CNN offers two significant advantages. Firstly, the AHT algorithms aid in the efficient and convenient re-training of the enhanced model by eliminating the need for hand-tuning hyperparameters. Secondly, as shown in Fig. 8, the enhanced model remains more accurate than the standard model when generalized to completely different mmWave communication scenarios, making it more reliable. In conclusion, the proposed model enhanced by the feature selection strategy produces higher accuracy and stronger generalization ability than the standard model. What’s more, we can see from Figs. 7 and 8 that the enhanced CNN outperforms the FCN in terms of both accuracy and generality as well, suggesting that CNN is a better choice. VI. CONCLUSIONIn this paper, we propose a generalized AHT-assisted CNN model for mmWave channel prediction under different communication situations. We integrate the AHT procedure into the training process to simultaneously save training effort and improve predictive performance. The diversification strategy is applied to enhancing the stability of the hyperparameter optimization process under different databases. Moreover, we use a novel feature selection strategy to make the model generalizable to different communication scenarios, which contributes to avoiding repetitive channel sounding. In parallel, coordinate transformation is utilized to reduce the sensitivity of the model to coordinate noise. Numerical results show that the enhanced model achieves desirable performance, including being convenient to train, less training data required, robust to coordinate noise, and strong generalization ability. Channel sounding in mmWave communication is of great importance for channel modeling but it is difficult in practical implementations. In this context, our proposed model can serve as an alternative approach to channel modeling for different communication systems with limited channel sounding overhead. BiographyChengfang YueChengfang Yue received the B.E. degree in Electrical and Electronic Engineering from University of Sheffield, UK, in 2015. He is currently working toward the Ph.D. degree in Control Science and Engineering at Wuhan University of Science and Technology. His current research interests include wireless communication technology and machine learning. BiographyHui TangHui Tang ui Tangui TangH received the B.E. degree in Electrical and Electronic Engineering and the Ph.D degree in Radio Physics from Wuhan University, China, in 2011 and 2016, respectively. She was a Visiting Ph.D student at Arizona state University, USA, from September 2014 to August 2016. She is currently an Associate Professor of Wuhan University of Science and Technology. Her research interests include machine learning, graph neural networks and radar signal processing. BiographyJun YangJun Yang received the B.E. degree in Computer Sci- ence and Technology and the Ph.D degree in Control Science and Engineering from Wuhan University of Science and Technology, China, in 2003 and 2016, respectively. She is currently an Associate Professor of Wuhan University of Science and Technology. Her research interests include machine learning, channel modeling and signal processing. BiographyLi ChaiLi Chai (S’00-M’03) received the BS degree in Applied Mathematics and the MS degree in Control Science and Engineering from Zhejiang University, China, in 1994 and 1997, respectively, and the PhD degree in Electrical Engineering from the Hong Kong University of Science and Technology, Hong Kong, in 2002. From August 2002 to December 2007, he was at Hangzhou Dianzi University, China. He worked as a Professor at Wuhan University of Science and Technology, China from 2008 to 2022. In August 2022, he joined Zhejiang University, China, where he is currently a Professor at the College of Control Science and Engineering. He has been a Postdoctoral Researcher or Visiting Scholar at Monash University, Newcastle University, Australia and Harvard University, USA. His research interests include distributed optimization, filter banks, graph signal processing, and networked control systems. Professor Chai is the recip- ient of the Distinguished Young Scholar of the National Science Foundation of China. He has published over 100 fully refereed papers in prestigious journals and leading conferences. He serves as the associate editor of IEEE Transactions on Circuit and Systems II: Express Briefs, Control and Decision and Journal of Image and Graphs. References
|