Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Abstract—This paper proposes a novel neural network archi- about the channel estimates and the hybrid precoding designs
tecture, that we call an auto-precoder, and a deep-learning based to significantly reduce the training overhead associated with
approach that jointly senses the millimeter wave (mmWave) the mmWave channel training and precoding design problem.
arXiv:1905.13212v1 [cs.IT] 30 May 2019
R = log2 I + R−1
Baseband Baseband H H H
Precoder Combiner
n W HFF H W , (2)
RF RF
NS N RF NS
FBB Precoder Combiner N RF WBB
FRF WRF
with F = FRF FBB and W = WRF WBB . The matrix
Nt Nr 1
Rn = SNR WH W represents the noise covariance matrix
RF Chain RF Chain
_
where SNR = NPSTσ2 , and with PT and σn2 denoting the total
+ n
transmit power and noise power.
Next, we assume that the RF beamforming/combining vec-
Fig. 1. A block diagram for the adopted hybrid analog/digital transceiver
architecture with a transmitter employing Nt antennas and NtRF RF chains
tors are selected from pre-defined quantized codebooks, i.e.,
and a receiver employing Nr antennas and NrRF RF chains. The precod- [FRF ]:,nt ∈ F , ∀nt and [WRF ]:,nr ∈ W, ∀nr . If the channel
ing/combining processing is divided between baseband precoding/combining is known, the hybrid precoder/combiner design problem can
FBB , WBB and RF precoding/combining FRF , WRF .
then be written as
Received Predicted
Sensing Transmit
Vector Beams
Predicted
Receive
Beams
Fig. 2. The proposed auto-precoder neural network consists of two sections: the channel encoder and the precoder. The channel encoder takes the vectorized
channel vector as an input and passes it through two complex-valued 1D-convolution layers that mimic the Kronecker product operation of the transmit and
receive measurement matrices in (13). The output of the channel encoder (the received sensing vector) is then inputted to the precoder network. The precoder
network consists of two fully connected layers and two output layers with the objective of predicting the indices of the NtRF RF beamforming vectors and
NrRF RF combining vectors of the hybrid architecture. The dimensions of the two output layers are equal to the sizes of the RF beamforming and combining
codebooks, F, W, i.e. the number of classes for the multi-label classification problem.
IV. MM WAVE AUTO -P RECODER : A N OVEL N EURAL rates as will be shown in Section V. Next, we briefly explain
N ETWORK FOR D IRECT H YBRID P RECODING D ESIGN the main concept of the proposed auto-precoder deep learning
model in Section IV-A and then provide a detailed description
In classical (non machine learning) signal processing, the of its two main stages, namely the channel sensing and hybrid
channel estimation and hybrid precoding design is normally beam prediction in Sections IV-B-IV-C.
done through three stages [2], [3]. First, leveraging the sparse
nature of the mmWave channels, the channel is sensed using
A. mmWave Auto-Precoder: The Main Concept
compressive measurements (that are normally random) [1], [3].
Then, the channel is reconstructed from the compressed mea- In this paper, we propose the novel auto-precoder neural
surements using, for example, basis pursuit algorithms. Finally, network architecture in Fig. 2. The auto-precoder neural net-
the constructed channel is used to design the hybrid precoding work consists of two main sections: (i) the channel encoder
matrices. The main drawback of this approach is that it does which learns how to optimize the compressive sensing vectors
not leverage the prior channel observations to reduce the to focus the sensing power on the most promising directions
training overhead associated with estimating the channels and and (ii) the precoder which learns how to predict the RF
designing the precoders. In this paper, we propose a novel beamforming/combining vectors of the hybrid architecture
neural network architecture that we call an ‘auto-precoder’ and directly from the receive sensing vector; i.e., the output of
a deep-learning approach that senses/compresses the channels the channel encoder. In order to achieve these objectives, the
and directly design the hybrid beamforming vectors from auto-precoder network is trained and used as follows.
the compressed measurements. Our approach achieves multi- • Auto-Precoder Training: In the training phase, the auto-
fold gains: (i) Different from the random measurements that precoder is trained end-to-end in a supervised manner.
are normally adopted in compressive channel estimation, our More specifically, a dataset of the mmWave channels and
approach learns how to optimize the measurement (compres- the corresponding RF beamforming/combining matrices
sive sensing) vectors based on the user distribution and the are constructed and the auto-precoder is trained to be
surrounding environment to focus the measurement/sensing able to predict the indices of the RF precoding/combining
power on the most promising spatial directions. (ii) The deep vectors of the hybrid architecture given the input channel
learning model learns (and memorizes) how to predict the vector. In this paper, we use the near-optimal Gram-
hybrid beamforming vectors directly from the compressed Schmidt hybrid beamforming algorithm in [11] to con-
measurements. This design approach significantly reduces the struct the RF beamforming/combining matrices. We note,
training overhead while achieving near-optimal achievable however, that other hybrid beamforming algorithms can
4
Predicted
Received
important to mention here that through the end-to-end Sensing
Vector
BS Beams
complex-valued neural network implementation of the con- The MTL strategy enables us to solve two related problems
volutional layers in [12]. using one neural network.
It is important to note here that the end-to-end training
of the auto-precoder (explained in Section IV-A) teaches D. Deep Learning Model Training and Prediction
this channel encoder in an unsupervised way to optimize
As briefly highlighted in Sections IV-A-IV-C, the proposed
its transmit/receive compressive channel measurement matri-
auto-precoder network in Fig. 2 is trained end-to-end as a
ces (or kernel weights). Intuitively, this optimization adapts
multi-task learning problem [13], which belongs to the super-
the measurement matrices to the surrounding environment
vised learning class. Essentially, we train the neural network
and user distribution and focuses the sensing power on the
based on a dataset of channel vectors and corresponding RF
promising spatial directions. After the model is trained, the
beamforming/combining vectors of the hybrid architecture.
kernels of the channel encoder will be directly employed
The target RF precoding/combining matrices are calculated
as the measurement matrices by the transmitter and receiver
based on the near-optimal Gram-Schmidt hybrid precoding
to sense the unknown channels. The output of this channel
algorithm in [11]. Following [14]–[17], the channel vectors
measurement (the receive sensing vector y) will be inputted to
are globally normalized such as the largest absolute value of
the precoder network in Fig. 2, which is the second section of
the channel elements equals one. The labels for the transmit
the auto-precoder, to predict the RF beamforming/combining
beamforming vectors (and similarly for the receive combining
vectors.
vectors) are modeled as NtRF -hot vectors, with ones at the
locations that correspond to the indices of the target RF
C. Hybrid beam predictions beamforming codewords (from the codebook F).
For the loss function, we use the binary cross entropy for
Given the received channel measurement vector y, we
the multi-label classification that corresponds to each task (i.e.,
leverage deep neural networks to learn the direct mapping
to the precoding and combining tasks). The total loss function
function from the received vector y and the beamform-
is the arithmetic mean of the binary cross entropies of the two
ing/combining vectors. For simplicity, we focus on predict-
tasks since they are equally important for the entire hybrid pre-
ing the RF beamforming/combining vectors FRF and WRF .
coding/combing design objective. To evaluation the prediction
Note that finding these RF beamforming/combining vectors
performance, we calculate the sample-wise accuracy of the
is what requires large training overhead in mmWave systems. (i)
predicted indices. Let ytrue denote the set of label indices and
Once the RF beamforming/combining vectors are designed, (i)
the low-dimensional effective channel WRF H
HFRF can be ypred denote the set of predicted indices. Then, the accuracy
easily estimated and used to construct the beseband precoders of sample i is defined as
(i)
and combiners. Further, since the RF beamforming/combining ytrue ∩ y (i)
pred
vectors are selected from the quantized codebooks F , W, we ai = (i) , (14)
ytrue
formulate the problem of predicting the indices of the RF
beamforming/combining vectors as a multi-label classification
where ∩ represents the intersection of two sets and · is
problem. Next, we briefly explain the adopted network archi- the cardinality of a set. Finally, the sample-wise accuracy is
tecture for the hybrid beam prediction. defined as
Network Architecture of ‘Precoder’: To predict the in- PN
dices of the RF beamforming/combining vectors from the i=1 ai
a= , (15)
receive measurement vector y, we propose the ‘precoder’ Nsamples
neural network architecture in Fig. 2. This network consists where Nsamples is number of samples. Note that since the
of two fully connected layers and two output layers, and number of labels for each channel is a fixed number for
it is fed by the output of the ‘channel encoder’ network both true labels and predicted labels, the two widely-adopted
described in Section IV-B. Each fully-connected layer is performance metrics in the multi-label classification problems,
followed by Relu activation and batch normalization. The namely precision and recall, reduce to (14).
network has two output layers; one predicts the indices of During this training process, the neural network architec-
the transmit beamforming vectors and the other one predicts ture in Fig. 2 achieves two joint objectives: (i) It optimizes
the indices of the receive combining vectors. The dimensions the transmitter/receiver measurement vectors (which are the
of the two layers are equal to the cardinalities of the transmit weights of the two convolutional layers in the channel encoder
and receive codebooks, F , W, i.e., the number of candidate network) in an unsupervised way to focus the sensing power
beamforming/combining vectors. Note that the number of on the most promising directions and (ii) it learns how to
candidate beams is also the number of classes for the multi- predict the RF beamforming/combining vectors of the hybrid
label classification problem. architecture directly from the channel measurement vectors
The auto-precoder network is trained end-to-end in a Multi- through the precoder network. Thanks to this design that
Task Learning (MTL) manner [13], by considering the design leverages the prior observations in designing the channel mea-
of RF precoding and combining matrices of the hybrid archi- surement/sensing vectors and hybrid precoding/combining ma-
tecture as two related tasks. This way, the network is trained to trices, the proposed approach significantly reduces the required
simultaneously optimize the two tasks which accounts for the training overhead to design the hybrid precoding/combining
dependence between the precoding and combining matrices. matrices, as will be shown in the following section.
6
TABLE I 25
Upper Bound with Perfect Channel Knowledge
T HE ADOPTED D EEP MIMO DATASET PARAMETERS
23 Proposed Direct Hybrid Precoding, Mt =Mr=2
DeepMIMO Dataset Parameters Value Proposed Direct Hybrid Precoding, Mt =Mr=4
Activate BS 4 Proposed Direct Hybrid Precoding, Mt =Mr=8
20
V. S IMULATION R ESULTS
5 10 15 20 25 30
In this section, we evaluate the performance of the proposed Total Transmit Power (dBm)
deep-learning based direct hybrid precoding design approach
using realistic 3D ray-tracing simulations. Fig. 4. The achievable rates of the proposed deep-leaning based joint channel
sensing and hybrid precoding approach for different values of total transmit
power. PT . These rates are also compared to the upper bound that designs the
A. Dataset and Training hybrid precoding/combining matrices assuming perfect channel knowledge.
This figure illustrates that the proposed solution can achieve near-optimal
Our dataset is generated using the publicly-available generic data rates while significantly reducing the training overhead, i.e. requiring a
DeepMIMO [18] dataset with the parameters summarized in small number of channel measurements, Mt , Mr ).
Table I. More specifically, we consider BS 4 in the street-level
outdoor scenario ‘O1’ communicating with the mobile users Fig. 4 illustrates the achievable rate of the proposed deep-
from row R1200 to R1500 with an uplink setup. For simplicity, learning based blind hybrid precoding approach and its upper
both the transmitter (mobile users) and receiver (base station) bound versus different values of total transmit power. Further,
are assumed to employ 64 antennas with 3 RF chains each. to draw some insights into the required training overhead,
For every user, we first construct the channel matrix using we plot the achievable rate of the proposed direct hybrid
the DeepMIMO dataset generator. Then, random noise is precoding approach for three different values of the channel
added to the channel matrix. The noise power is calculated measurements, Mt = Mr = 2, 4, 8. This figure adopts
based on a bandwidth of 0.5GHz and receive noise figure the system and channel models described in Section V-A
of 5dB. The channel is also adopted to construct the target where both the transmitter (mobile user) and receiver (base
RF precoding/combining matrices of the hybrid architecture station) employ a hybrid transceiver architecture with 64-
using the near-optimal Gram-Schmidt based hybrid precoding element ULA and 3 RF chains. First, Fig. 4 shows that the
algorithm in [11]. the pair of the noisy channel and the corre- performance of the proposed deep-learning based direct/blind
sponding codebook indices of the RF precoders/combiners is hybrid precoding solution approaches the upper bound (which
then considered as one data point in the dataset. assumes perfect channel knowledge) at reasonable values of
The constructed dataset is used to train the proposed auto- the transmit power. Further, and more interestingly, this figure
precoder neural network adopting the cross-entropy based loss illustrates the significant reduction in the required training
function defined in Section IV-D. Our model is implemented overhead compared to classical (non-machine learning) solu-
using the Keras libraries [19] with a Theano [20] backend. We tions. For example, with only 4 channel measurements at both
use the Adam optimizer with momentum 0.5, a batch size of the transmitter and receiver, i.e., a total of 4 × 4 = 16 pilots,
512, and a 0.005 learning rate. the proposed deep-learning based channel sensing/precoding
solution achieves nearly the same spectral efficiency of the
B. Achievable Rates upper bound. This is instead of the 64 × 64 = 4096 pilots
To evaluate our deep-learning based direct hybrid precoding required for exhaustive search and almost one tenth of that,
solution, we calculate the achievable rate using the predicted ∼ 300 − 400, pilots needed by compressive sensing solutions
precoding/combining indices and compare it with the optimal [6]. This significant reduction in the training overhead
rate. More specifically, for a given noisy channel measure- is thanks to the proposed deep learning model which
ment, we use the proposed auto-precoder neural network leverages the prior observations to optimize the channel
model in Section IV to predict the indices of the RF pre- sensing based on the environment/user locations and to
coding/combining matrices of the hybrid architecture. Then, learn the mapping from the receive channel sensing vector
adopting the baseband precoding/combining design in (9)-(10) and the optimal precoders/combiners.
we calculate the achievable rate as defined in (2). This rate is
then compared to the optimal rate that is achieved when the C. Sample-Wise Accuracy
precoding/combining matrices are calculated as described in In addition to the achievable rate (which is the main
Section V-A with perfect channel knowledge. objective of this work), we also evaluate the performance
7