1. Introduction
Distributed transmit beamforming with
N cooperating transmitters can provide received power
-fold larger than that of a single transmitter. Over the past decade, there has been significant progress in demonstrating the feasibility of attaining the precise carrier frequency and phase synchronization required to realize these gains [
1,
2,
3,
4]. In this paper, we build on these advances for design of a concept system that we term distributed base station (DBS), targeting significant improvements in communication link range and/or data rate. As shown in
Figure 1, a DBS comprises
N opportunistically placed, low-cost, transmitter nodes, without wired connections between the nodes. Our goal is to leverage the
-fold received power gain provided by distributed transmit beamforming to significantly enhance downlink range and/or spectral efficiency. While a DBS can be employed to enhance communication in existing WiFi and LTE bands, the approach is particularly interesting for white space frequencies (e.g., 50–800 MHz). These frequencies propagate well and are therefore well matched to long-range applications such as rural broadband. However, good propagation also leads to poor spatial reuse when employing omnidirectional transmission. Beamforming using multiple antennas can enhance spatial reuse, but centralized antenna arrays are bulky at large wavelengths. DBS allows the use of low-cost transmit nodes with moderate transmit power to emulate a powerful transmitter with a highly directional steerable antenna.
For
N coordinating transmitters, we can obtain an
N-fold power pooling gain even without beamforming. In practice, this would require loose timing coordination between the transmitters (e.g., so the delay spread seen by the receiver across transmitters is smaller than the size of an OFDM cyclic prefix). While this addition is noncoherent, and thus leads to fading on narrowband channels, this effect is alleviated by frequency diversity in wideband systems. Thus, we envision a system which starts with an
N-fold power pooling gain and then adapts to higher spectral efficiency as the DBS trains its beam towards the desired receiver, with the maximum spectral efficiency corresponding to the
-fold ideal beamforming gain. As an example of the system enhancements possible with a 10-node DBS (our running example), consider a rural broadband link served by a single transmitter node in white space frequencies (50–800 MHz). For a receiver at the cell edge, which could only sustain a very low rate control channel at
dB SNR with a single base station transmitter, a 10-node DBS boosts the SNR at the receiver to 5 dB with power pooling, and to 15 dB with ideal beamforming. For typical SNR versus modulation and coding scheme (MCS) values for LTE adaptive modulation [
5], we can sustain QPSK with rate
coding with power pooling and 64QAM with rate
coding with beamforming. Thus, for a channel bandwidth of 20 MHz, a DBS makes it possible to provide broadband data rates of 20–80 Mbps at a range where a single transmitter system could barely establish a control link. Note, however, that in order to achieve these gains, the distributed beamforming scheme must operate effectively in regimes where the received SNR for any given transmit node is low. Thus, we cannot use, for example, the well-known one-bit feedback based beamforming strategy originally proposed in [
2], which has been shown to be fundamentally limited in its performance for low per-node SNR [
6].
In terms of the range extension possible with an
N-node DBS, a free space propagation model with inverse square decay of power with distance would predict an
N-fold range increase, corresponding to the
-fold power gain due to beamforming. The gains when we use more realistic propagation models are smaller, but still significant. The power of the DBS concept lies in our ability to harness MIMO at lower carrier frequencies. As an example link budget using the Hata propagation model in
Section 6.1 indicates, by using a white space frequency of 200 MHz, we can attain spectral efficiencies larger than 3 bps/Hz at a range of 14 km, using a 10-node DBS with a relatively small transmit power per node of 20 dBm (the emitted power of a typical WiFi node). If we now use a larger carrier frequency of 800 MHz, the range drops to approximately 6 km.
Contributions: We propose and evaluate a wideband system design that addresses the key technical hurdles for realizing the DBS concept. We consider an FDD system in which the receiver sends feedback that enables the transmitters in the DBS to synchronize. In order to attain protocol-level scalability, we constrain the feedback to be aggregate (i.e., not directed at any specific transmit node), so that the receiver is oblivious to the number and identity of the transmit nodes. The key contributions are as follows.
- (1)
We design a feedback-based synchronization strategy, which we term deterministic orthogonal sequence training (DOST). We explore its properties in a narrowband setting, comparing it against a number of previously proposed strategies to show that it is better matched to the low per-node SNR regime of interest to us. We also show that DOST is resilient to severe quantization on the feedback link, which as we discuss shortly, can become a bottleneck in the DBS system.
- (2)
We show that DOST extends naturally to wideband frequency-selective channels by considering an OFDM system in which training sequences are sent on a designated set of pilot subcarriers. The receiver feeds back a corresponding sequence of (quantized) received complex amplitudes, which the transmitters use to estimate their channels on the pilot subcarriers. These are then interpolated across subcarriers by each transmitter and used to beamform on the data subcarriers. Assuming a feedback delay which is smaller than the time constant of channel variations, this approach scales to arbitrarily low per-node SNRs: measurement noise is averaged out effectively over a period scaling inversely with the per-node SNR.
- (3)
We compactly characterize the downlink performance for a DBS via an outage capacity analysis for a well-accepted 3GPP channel model.
- (4)
We note that feedback can become a bottleneck to varying degrees, with the rate of channel time variations that can be supported depending on the sophistication of the uplink reception strategy. We use an outage capacity analysis to compare the channel coherence times that the DBS can support with and without distributed reception on the uplink.
Organization: We put our work in context by briefly summarizing related work in
Section 2. In
Section 3, we describe the overall system design and provide some example results. In
Section 4, we describe the DOST scheme tor feedback-based training and provide performance evaluations for a single subcarrier, including the effect of drastic feedback quantization. We justify our design choices by comparing DOST with alternative techniques such as one-bit feedback-based beamforming [
7] and time-multiplexed training [
8]. In
Section 5, we provide results showing the performance of DOST for a wideband system, showing that simple interpolation across the pilot subcarriers to estimate the channel for the data subcarriers works well.
Section 6 provides an outage capacity analysis that compactly characterizes the coded performance of the wideband system, both for the downlink and for the feedback on the uplink.
Section 7 contains our conclusions.
2. Related Work
With the increasing availability of low-cost front-end elements, architectures that take advantage of the degrees of freedom offered by massive deployment of antenna elements have gained in popularity. Acquisition of adequate channel state information at the transmitter (CSIT) is crucial for FDD massive MIMO systems and has been of particular interest in literature with traditional codebooks for channel feedback [
9,
10], which requires the number of feedback bits to scale linearly with the number of BS antennas [
11]. Efficient codebook design based on the channel statistics [
12] and sparsity inspired approaches are proposed in [
13,
14] to reduce feedback overhead. The fundamental differences between this body of work and our framework are as follows. First, in order for the network protocols to scale with the number of distributed transmitters, and to allow opportunistic expansion of the DBS, we constrain the receiver to be oblivious to the number and identity of transmitters. Thus, instead of performing channel estimation and then producing quantized feedback, the feedback must be based on the receiver’s
aggregate measurements. This still allows us to consider standard training strategies, in which different transmitters send orthogonal training sequences, but constrains the form of feedback the receiver can send back. This implies, for example, that the receiver cannot perform spatial channel estimation followed by codebook-based quantization, or exploit sparsity, unlike in existing feedback-based techniques in massive MIMO. Second, the impact of operating in the low per-node SNR regime has not been considered in prior work on massive MIMO feedback. As shown in this paper, this limitation on the feedback rate can fundamentally limit the channel coherence times that can be supported.
A simple example of distributed beamforming with aggregated feedback is the one bit feedback algorithm [
2,
7], in which transmitters use small random phase perturbations to perform stochastic ascent on the received signal strength, based on the receiver feedback, broadcast to all transmitters, of one bit per iteration. This approach is simple and has formed the basis for several prototypes [
3,
15]. The relatively slow convergence of the original algorithm (e.g.,
iterations to reach 75% of ideal beamforming gain [
2]) can be improved to a limited extent by strategies such as exploiting knowledge of previous perturbations [
16], or using knowledge of channel statistics at the receiver [
17]. However, the method is fundamentally mismatched to low per-node SNRs [
6], roughly speaking, because noise masks the effect of small phase perturbations. This motivates our approach, in which the transmitters employ large phase perturbations during a training phase, rather than attempting to make small adjustments while beamforming.
The one-bit feedback algorithm has been extended to wideband regimes in a prior paper by the authors [
18], by adding an additional bit per subcarrier which enables enforcement of phase continuity across subcarriers. This is fundamentally different from the training-based approach in this paper, which enables explicit channel estimation on pilot subcarriers for each transmitter, and thus is amenable to standard interpolation across subcarriers.
Once we commit to a training phase, one possible approach is for the transmitters to take turns transmitting in training slots, with the receiver sending (quantized) feedback corresponding to each slot. Such time-multiplexed training has been successfully prototyped [
19], and has been studied with quantized feedback in [
8]. The algorithm is energy efficient, as only one node is active per iteration, but unlike the DOST scheme proposed in this paper, it does not utilize integration over time to combat noise, so that its performance suffers in the low per-node SNR regime, as we show in our numerical results. The time-multiplexed approach also does not scale well at the protocol level because of the dependence of the training frame structure on the number of transmitter nodes and the coordination required between them to take turns.
Our work, and the related work discussed above, falls into the category of all-wireless distributed beamforming with explicit feedback, which allows flexible deployment of DBS supporting FDD operation. It is worth mentioning recent work on all-wireless distributed beamforming based on channel reciprocity that relies on tight pre-synchronization of the cooperating nodes to emulate a centralized array with a common time base [
20,
21,
22]. Finally, there is also significant recent work on distributed MIMO based on coordination of infrastructure nodes (WiFi access points or cellular base stations) via a fast wired backhaul [
23,
24,
25,
26]. Our emphasis is to scale up the number of nodes without fast wired backhaul and achieve massive MIMO gains [
27,
28,
29].
The present paper builds on our prior conference paper [
30], which introduced the DOST scheme and also discussed extensions to wideband systems. However, it goes well beyond [
30] by presenting a DBS concept system built around DOST, including an OFDM system design roughly consistent with LTE, specific prescriptions for pilot design, and outage capacity analyses for downlink and uplink which compactly characterize system-level performance.
3. System Model
The nodes in the DBS depicted in
Figure 1 cooperate to send a common message to a distant receiver over a noisy multipath channel. We consider OFDM with a set of subcarriers designated as pilots. The receiver broadcasts explicit aggregate feedback on the complex signals received on the pilot subcarriers. Each transmitter in the DBS uses this feedback to estimate its complex channel gains to the receiver on the pilot subcarriers, and interpolates these to estimate the channel gains on the data subcarriers. Each transmitter then adjusts its phase on each subcarrier to compensate for the channel phase, in order to align coherently at the receiver.
The transmitters are assumed to be synchronized in time and frequency, and the channel is assumed to be quasi-static (i.e., it can be modeled as time-invariant with respect to the time constant of the feedback-based channel estimation strategy). Timing synchronization is relatively straightforward with the OFDM system model because timing offsets smaller than the length of the cyclic prefix are tolerable. As long as each transmitter maintains a fixed framing with respect to its own clock, drifts between the transmitters’ clocks can be compensated for by the feedback-based phase synchronization. Frequency synchronization can be achieved by each node synchronizing to the receiver node (as done in prototypes such as in [
3,
31]) or to a master node within the DBS, and the effect of small residual frequency offsets is similarly compensated for by the phase compensation algorithm.
We now describe the signal and channel models employed in the paper. We discuss and justify our design choices in more detail in later sections of the paper.
3.1. Signal Model
We denote the channel from node
i to the receiver on subcarrier
k by the complex gain
, and the receiver’s phase offset relative to transmitter
i by
. Transmitter
i applies phase control via a beamforming weight of
on its
kth subcarrier. The received signal, after multiplying by the conjugate of the unit-amplitude pilot symbol, is given by
The corresponding normalized received signal strength (RSS) is given by
and is used as a performance metric to compare different beamforming algorithms.
The aim of distributed transmit beamforming is to maximize RSS. This is achieved when each transmitter chooses a beamforming phase that reverses its total offset relative to the receiver allowing all signals to combine coherently upon reception. The optimal solution is therefore
up to a common constant shift across all nodes. The received RSS is then equal to
and the normalized RSS
approaches the maximum achievable value.
3.2. Channel Model
The multipath channel between a typical transmitter node and the receiver is modeled as
where
denotes the number of paths,
the delay, and
the complex gain of path
p. For concreteness, we consider Rayleigh fading on each path, setting
where
are i.i.d. complex Gaussian with distribution
and
is the normalized square root of the power delay profile (PDP) such that
.
The frequency response for such a channel is
For each frequency , is zero-mean complex Gaussian with variance . Thus, the channel responses of each transmitter at different frequencies are identically distributed but correlated random variables with distribution .
In our numerical results, we use the 3GPP Extended Pedestrian A (EPA) channel model parameters shown in
Table 1 to generate channels for each node in the DBS. Different nodes therefore have the same power-delay profile, but different channel realizations corresponding to i.i.d. draws of the
. We note that this channel model is not intended to provide a physical model of multipath components, but rather, may be viewed as a non-uniform tapped delay line representation of a band-limited channel. We have also considered dithered versions of the delays for different nodes, and verified by simulations that the channel statistics in frequency domain do not change. Thus, the channel model should be viewed as a non-uniform tapped delay line representation, rather than a model for physical multipath components.
4. Per-Subcarrier Feedback-Based Distributed Beamforming
In this section, we describe and justify our choice of training strategy through numerical comparison against alternative approaches. We focus on a narrowband system which serves as a model for a single pilot subcarrier in the wideband OFDM framework. As mentioned, we are interested in techniques that scale well, in terms of both performance and protocol simplicity, as the number of transmitter nodes increases and as the received SNR per node approaches zero.
4.1. Deterministic Orthogonal Sequence Training
In this scheme, each node uses a predefined sequence of beamforming weights over L training transmissions and the L complex gains measured by the receiver are quantized and broadcast to the transmitters. By using orthogonal or quasi-orthogonal weight sequences on different nodes, each node can extract its channel from the feedback independently from other nodes.
Consider the
training matrix
, the
i’th column of which is the weight sequence used by node
i. To design
N orthogonal sequences, each sequence must be at least of length
N, meaning the training period, which is equivalent to convergence time for the iterative approaches, is equal to
N and scales linearly with array size. Using orthogonal sequences is then equivalent to choosing a training matrix for which
is diagonal. This orthogonality requires some coordination between transmitters to ensure one-to-one assignment of sequences to nodes. This requirement can be relaxed by using quasi-orthogonal pseudorandom sequences that nodes generate independently; the normalized interference between sequences gets attenuated as their length grows. While it is possible to employ completely uncoordinated training by the latter choice, in practice, the coordination required for implementing truly orthogonal sequences (which provide the best possible performance for a given training duration and power) is minimal. There are many possible choices of training sequences, but for concreteness, we consider the DFT matrix in our results:
This is because DFT sequences are not only orthogonal, but they remain orthogonal when cyclically shifted by any amount. Thus, a transmitter can use any L-sized block of feedback to estimate its channel, without incurring interference from the sequences sent by the other transmitters. A receiver can therefore snoop on the pilot subcarriers at any time, and generate feedback for the transmitter nodes in the DBS. Similarly, any transmitter node can join or leave the DBS at any time, assuming basic OFDM frame alignment and frequency synchronization is maintained. This makes deployment and operation particularly simple.
The observations at the receiver, collected over times
, can be written as the
vector
where
is the
channel vector across different transmitters and
is the receiver noise.
The least squares estimate for the channel vector is given by
assuming that
and
has rank
N. Each node can thus obtain its channel estimate by taking the inner product of its corresponding row in the matrix
and the channel measurement feedback vector.
The Cramer–Rao lower bound on error covariance is
. For each transmitter, the error covariance is bounded as
with the bound attained for orthogonal training (diagonal
). In this case, each node can estimate its channel by separately correlating the observations with its own training sequence:
where
is the
nth column of the training matrix. The estimation error covariance
can be made arbitrarily small by increasing the training interval
L. This also demonstrates the power-pooling advantage of the DOST algorithm; with
L scaling linearly with
N, the estimation accuracy improves as array size grows and longer links with lower RSS can be supported.
Quantization
In practice, the complex received signal amplitude measured at the receiver must be quantized to a limited number of bits and broadcast by the receiver. The variance of the received complex amplitude scales as N (the transmitted signals add up incoherently during the training period), thus a natural question is whether the quantization resolution also needs to be enhanced as N increases. Fortunately, the answer is no: as long as the receiver scales its quantizer step size as to accommodate the amplitudes it is seeing, we can use a fixed number of quantization bins, and average out the quantization noise across the training period.
The channel estimate at transmitter
n with quantized feedback can be written as
where
is the quantization noise vector. Assuming quantization noise is distributed uniformly over the span associated to each level, the variance of any element
of the quantization noise vector scales as
If the quantization noise values can be approximated as independent over time, we have
so that the effect of quantization noise on channel estimation can be made independent of
N by scaling
L linearly with
N. Thus, we can use a fixed feedback rate even as we increase the number of transmitters
N, as long as the length of the training period scales linearly with
N.
4.2. Alternative Strategies
The goal here is not to be comprehensive, but to show that the proposed DOST strategy is better matched to the low-SNR regime than methods that have been suggested in the literature, and natural variants thereof. To this end, we consider per-node deterministic training, the one-bit feedback algorithm, and variants of the latter that employ 2 bits per iteration.
4.2.1. Successive Deterministic Distributed Beamforming
A special case of deterministic orthogonal training is time-multiplexed training where only one transmitter is active at a time and the phase offset measured at the receiver is fed back to individual nodes successively. This procedure, termed Successive Deterministic Distributed Beamforming (SDDB) in [
8], corresponds to setting the training matrix to identity, i.e.,
. As our analysis and numerical results demonstrate, this method is poorly matched to the low per-node SNR regime of interest to us, as a transmitter is not able to use the entire training period to average out noise. Of course, if the per-node SNR is large enough, SDDB may be preferable in terms of power conservation as only one transmitter is active at any time during a training period of similar length. This may be the case for shorter range applications, in which the goal of distributed beamforming is to reduce transmitted power rather than to obtain range extension for a given transmitted power. On the other hand, SDDB is more resilient to quantization than DOST, as the dynamic range of the received signal is smaller when a single node is transmitting at a time.
4.2.2. One-Bit Feedback
The one-bit feedback algorithm (OBF) is a simple randomized iterative procedure that aims to synchronize the phase of transmitters using a single bit of feedback for each measurement. This feedback is aggregated for all transmitters instead of explicitly estimating CSI for each node. The algorithm proceeds as follows. In each iteration, transmitters perturb their phase randomly and independently from a predefined distribution and transmit a common signal. The receiver measures the combined response and broadcasts a single bit of feedback. The feedback bit
that is sent at the end of time slot
t indicates the result of comparison of the current measured RSS with the highest RSS observed over a moving window of the last
M iterations:
where
Each transmitter updates its phase according to the feedback from the receiver by keeping the random perturbation for a positive feedback and discarding it in the case of negative feedback:
where
is the phase perturbation added in iteration
t.
4.2.3. Randomized Two Bit Feedback Algorithm (R2BF)
The authors of [
17] propose a modified version of OBF, namely the randomized 2 bit feedback algorithm (R2BF), in order to speed up convergence. Assuming that the receiver has knowledge of the maximum possible RSS value obtained by perfect beamforming, the feedback bits are set as follows:
and the phase update (
8) becomes
In the next time slot,
is chosen from a different distribution depending on the feedback, i.e.,
where
where
. This approach increases the convergence speed of 1BF by approximately 25%, but with the additional requirement of the receiver knowing the maximum RSS, which in turn requires knowledge of the number of transmitters and the channel statistics. This method therefore requires a higher level of coordination between nodes and is less robust and distributed.
4.2.4. Modified Two Bit Feedback Algorithm (M2BF)
We propose a different modification of one bit feedback, where the additional bit of feedback is used to quantify the
amount of improvement obtained from the perturbations. The additional feedback bit relative to Equation (
6) is generated as follows:
where
and
are predefined constants dependent on channel statistics, but independent of
N. If the RSS
improvement from current random phase perturbations is above threshold
, all transmitters make use of this knowledge and apply the previous perturbations again in the next iteration. If the
degradation caused by the perturbation is more than threshold
, the phases are reversed and transmitters perturb their phases in the opposite direction. Therefore,
becomes dependent on the previous perturbation
as follows:
4.3. Numerical Results and Comparisons
To evaluate and compare the beamforming performance and convergence speed of the preceding algorithms, we fix the number of feedback bits to 2 and plot the progression of each algorithm with the number of iterations. We compare DOST with 2-bit quantized feedback against the 2-bit SDDB, M2BF, and R2BF strategies discussed above. While our later system-level numerical results are for a DBS with 10 nodes, in this section, we consider a larger number of nodes () in order to stress test the feedback strategies considered. We investigate the evolution of beamforming gain as a function of iterations using Monte Carlo simulations.
The distributed beamforming schemes used in the simulations have the following parameters: the phase perturbations
are generated from the uniform distribution
. R2BF uses
,
,
, and constant thresholds of
,
to decide from the
where RSS fits in (
9). For M2BF,
and the phase perturbations are designed to decay exponentially from
with the number of iterations.
We first focus on understanding the effect of quantization in a noiseless setting.
Figure 2 shows, at each iteration, the RSS level that would be obtained by nodes using their current channel estimate for beamforming. The curves of
Figure 2 are the result of averaging over 2000 realizations of an
element array in a noiseless setting. For deterministic algorithms, training is stopped after
iterations, which constitutes one “batch” of training. In the absence of noise, the stochastic algorithms, R2BF and M2BF, converge asymptotically to optimal beamforming, while deterministic algorithms hit an performance gap of 1 and 2 dB away from optimal beamforming for SDDB and DOST, respectively, due to feedback quantization. When the feedback link is not a bottleneck, increasing the number of feedback bits can also be used to decrease quantization loss, but our interest is in the low per-node SNR regime, where this is not a feasible strategy. Thus, in a noiseless setting with severe feedback quantization, the one-bit feedback algorithm and its variants actually perform better than deterministic training. Moreover, among the deterministic strategies, time-multiplexing across nodes as in SDDB is better than DOST, as the dynamic range of the received signal is smaller.
However, the picture is quite different when we consider the low per-node SNR regime of interest to us.
Figure 3 shows the evolution of the beamforming algorithms at per-node SNR of
dB. As the SDDB scheme does not get the benefit of time averaging, it falls 7 dB short of the ideal 20 dB beamforming gain after
N iterations, whereas DOST comes to within 2 dB of the ideal beamforming gain. The one-bit beamforming schemes perform very poorly and we do not plot it. Among its two-bit variants, the M2BF scheme performs better than the R2BF scheme, but falls well short of the ideal beamforming gain: 8 dB lower after
iterations, and 5 dB lower even after 500 iterations. Thus, the DOST algorithm is by far the most resilient at low per-node SNR.
Impact of Quantization
We now explore the impact of quantization further, by varying the number of bits of feedback per iteration in the DOST and SDDB algorithms. The number of feedback bits is fixed to 2 for M2BF and R2BF, thus we do not consider those schemes here.
Figure 4 shows the performance of the two algorithms with different levels of feedback quantization for different per-node SNR after 100 iterations of training. We consider 2-bit feedback quantization to quantize both real and imaginary parts of the received baseband signal. With 2-bit quantization, DOST is able to achieve within 2 dB of the ideal solution. Increasing the number of quantization bits to 4 bits improves both algorithms by approximately 1 dB and further increasing it to 6 bits gives very slight performance improvement. These results show that DOST can achieve near-optimal beamforming gains with heavily quantized feedback, as low as 2 bits per iteration, making it competitive with stochastic ascent approaches like R2BF and M2BF, even in noise-free conditions where they perform best.
5. OFDM Pilot Design
To extend the framework to wideband, an OFDM framework is utilized wherein DOST is applied on a subset of the OFDM subcarriers by placing training pilots at known positions in the OFDM symbol grid. Different pilot placements are possible for the training, including the block type, the comb type, or 2D-grid type [
33]. In a block type arrangement, the pilots are placed on all subcarriers in a few OFDM symbols; in the comb type, the pilots are present in all OFDM symbols over a subset of subcarriers as shown in
Figure 5; and in the 2D-grid type, the pilots are present in a subset of OFDM symbols over a subset of subcarriers. Therefore, the number of pilots in the 2D-grid pattern are less than the block and comb type pilot arrangements. Our goal here is to learn the channel as quickly as possible; thus, for any given subcarrier, it is best to concentrate our pilot resources in time (over
successive OFDM symbols for an
N-node DBS) so as to get the required feedback from the receiver as quickly as possible. This is particularly important for maximizing the rate of channel time variations a DBS can support, because of the relatively low rate of feedback available on the uplink (see
Section 6). However, by exploiting the continuity of the channel across frequency, we only need to employ pilots for a subset of subcarriers, and estimate the optimum beamforming weight for all other subcarriers via interpolation in the frequency domain. We therefore consider the comb type pilot arrangement shown in
Figure 5 for the DBS deployment.
A number of different interpolation strategies can be used to extend the pilot subcarrier channel estimates to the remaining subcarriers, including linear interpolation, second order interpolation, low pass interpolation, spline cubic interpolation, and time domain interpolation. We consider low-pass interpolation, which has been shown to work better with comb type pilots [
33].
The typical LTE system parameters shown in
Table 2 are used in the simulations, and comb type pilots are placed at all OFDM symbols on a subset of 200 equispaced subcarriers. The minimum required number of OFDM symbols for training is
, and the minimum required time for training is 0.71 ms for
nodes. Note that a standard 2D grid type pilot arrangement, as illustrated in
Figure 6, would require a longer training time. For the system parameters of
Table 2, the minimum required number of OFDM symbols for training is also
, however, a subset of OFDM symbols are used as the pilots and the required time for training is
OFDM symbols which corresponds to a minimum training time of 5 ms for
nodes. This gap grows linearly with the number of nodes and can become a bottleneck when scaling to larger arrays.
Numerical Results
Simulation results using uncorrelated EPA channels with the power delay profile of
Table 1 and system parameters of
Table 2 are reported here for the wideband setting. The comb type pilot arrangement is used with subcarrier spacing of 6.
Figure 7 shows the performance of beamforming algorithms averaged over subcarriers for
nodes. DOST achieves within 3 dB of the ideal solution, and works better than SDDB at low per-node SNR. The degradation due to interpolation of comb type pilots is only 0.5 dB compared to a setting in which all subcarriers are used as pilots.
6. System Performance
We now provide performance characterization via outage rates for the downlink transmission of pilots and data and for feedback broadcast in the uplink direction. For simplicity, we use the terms “capacity” and “outage rate/capacity” to denote spectral efficiencies, either for a narrowband system (modeling a single subcarrier), or averaged over subcarriers for a wideband system. We use the term “data rate” when we multiply such spectral efficiencies by the bandwidth.
The forward link enjoys the benefits of
N-fold power pooling gain during training, and
-fold distributed transmit beamforming gain during post-training data transmission. The rate of the feedback link depends on the sophistication of its reception strategy, as well as its allocated resources, which may be less than that of the forward link. There is no power pooling on the uplink and transmission is at an
N-fold disadvantage in this direction relative to downlink. This asymmetry may be offset by using distributed receive beamforming in the feedback direction to pool the resources of the array and allow uplink scaling to keep up with downlink. Different analog and digital receive beamforming algorithms have been proposed in previous works such as in [
4,
34]. In the worst case, however, feedback is delivered over a SISO channel, either by having each node decode and use the feedback independently or use a single designated node for feedback reception. In this case, if the downlink power emitted by a single node in the DBS cluster and the uplink power emitted by the user node are comparable, then the feedback link may well be the scaling bottleneck and necessitate longer symbol durations to build up SNR, which will limit the rate of channel time variations that can be supported by the distributed array.
Figure 8 illustrates an example frame structure. During the startup phase, data could be sent in power pooling mode at lower spectral efficiency, while sending training on the pilot subcarriers. Beamforming can be applied once the feedback corresponding to the pilots has been received. Once continuous communication has been established, feedback regarding the designated set of pilot subcarriers on the downlink in frame
f is sent back during the next frame
, and the results are applied for distributed transmit beamforming in frame
. With such a scheme, for the first and second frames, data could still be sent in power pooling mode at lower spectral efficiency, with distributed beamforming enabled from the third frame onward. For continuous communication, we would be beamforming in a given frame based on pilots sent two frames back. If the feedback link is the bottleneck, then the frame length
must be long enough to carry back the feedback corresponding to
pilot subcarriers on
OFDM symbols, which amounts to
bits using 2-bit feedback. For feedback of rate
, this corresponds to a frame length of
, and a channel coherence time of
. If the feedback link is not the bottleneck, then the minimum frame length can be set to
, where
is the OFDM symbol length, which is governed by the channel delay spread and the overhead allowed for the cyclic prefix. We provide numerical values for our running example of a 20 MHz downlink over an EPA channel model at the end of this section.
We first discuss achievable performance on the downlink, and then consider the feedback link. In both cases, we use outage rates for a compact bottom-line characterization.
6.1. Downlink Performance
We assume that each node in the DBS cluster applies a phase correction on each subcarrier based on its channel estimate, and employs a uniform power distribution across subcarriers. As the DBS nodes do not have information regarding the relative channel strengths across subcarriers or transmitters, more optimal power allocation methods such as water-filling are not possible. Even for a narrowband channel, using multiple transmitters for distributed beamforming provides spatial diversity, thus we derive a pessimistic estimate of outage spectral efficiency by ignoring frequency diversity and focusing on a single subcarrier at frequency . The channel seen by node i is denoted as . Upon ideal phase compensation, the net channel seen at the receiver is given by , where is the vector of channel gains corresponding to the N nodes in the DBS cluster. Modeling the channels as zero mean complex Gaussian normalized as , the effective channel amplitude gain is a sum of i.i.d. Rayleigh random variables, each with mean squared value of one.
Assuming that each transmitter applies power
P to each subcarrier, the outage probability for a narrowband system operating at
is given by
The -outage capacity is the maximum rate R such that is less than .
Letting
denote the CDF of
, we see that
As
is a sum of i.i.d. random variables, we get insight, and a good approximation, by applying the central limit theorem. That is, we can approximate
as Gaussian with mean
and variance
. Using this approximation in (
13), we obtain that
where
denotes the complementary CDF of a standard Gaussian random variable. This indicates that that the outage capacity shows a
growth with the number of nodes, with
backoff within the argument of the logarithm in order to handle the tails.
The Gaussian approximation works well for moderately large
N, including our running example of
, and provides insight into the benefits of both spatial diversity and beamforming. We note, however, that for small
N, the outage capacity approximation can be improved by using a small argument approximation to the CDF
F of a sum of i.i.d. Rayleigh random variables [
35], given by
where
is a normalized argument for the CDF. This approximation, when used in (
13), is excellent for small values of
t which is the regime of interest for the outage probability
.
We compare these approximations with simulations in the next section.
Numerical Results
Figure 9 shows the ergodic capacity and the outage rate versus the number of transmitters at
dB SNR per node for a narrowband channel with ideal channel state information. The ergodic capacity and the
outage rate curves are obtained with Monte Carlo simulations. The analytical outage capacity approximation for sum of Rayleigh random variables in (
15) matches Monte Carlo simulations very well and the Gaussian approximation of the sum of Rayleigh random variables (
14) is slightly pessimistic for the small number of nodes. The difference between ergodic capacity and outage rate diminishes as the number of nodes increases because the diversity gain provided by multiple nodes reduces the variance of the aggregate channel and, in turn, the variance of spectral efficiency. It can be observed that, with
nodes, the outage capacity of
bps/Hz can be obtained at
dB SNR per node.
Figure 10 shows Monte Carlo simulation results for outage capacity versus number of transmitters applied to the wideband setting (i.e., where the spectral efficiency is averaged over the signal bandwidth) with parameters in
Table 2 at
dB average SNR. The ideal CSI curve shows the capacity when the channel is known to all nodes and perfect beamforming is applied over the entire frequency band. The DOST curve shows Monte Carlo simulation results with 2 bits of feedback per pilot subcarrier. The heavily quantized DOST algorithm provides significant gains in terms of capacity and is able to achieve outage rate of
bps/Hz using 10 nodes. Thus, even while operating at a per-node SNR of
dB, DOST can yield a data rate of about 50 Mbps over a 20 MHz band, after accounting for the overhead of reserving
of the subcarriers for pilots, under the assumption that we would like to be as reactive to channel time variations as possible and therefore insert comb type pilots into every OFDM symbol. For transmit power of 20 dBm (100 mW) per DBS node, isotropic antennas, and receiver noise figure of 6 dB, the attainable range using the Hata propagation model at 800 MHz carrier frequency with 30 m DBS node height and 1.5 m receiver height is about 6.6 km, allowing for a 5 dB implementation margin (we already account for fading in our formulation, thus we do not require excess link margin to accommodate it). The range that can be attained in the same setting for the same target data rate is 2.3 km for a single node and 4.2 km with power pooling. As expected, the corresponding numbers at a lower carrier frequency of 200 MHz are better: 5 km for a single node, 9 km with power pooling, and 14.3 km with DOST.
6.2. Feedback Link
While we do not consider detailed design of the feedback link, we provide insight into the impact of reception strategy by comparing three different options. The first and simplest approach is to designate a single node (e.g., one of the DBS nodes) as receiver for the feedback. In this case, the received SNR is very low (e.g.,
dB for our running example, assuming that the emitted power from a DBS and user node are similar), and the feedback link becomes a significant bottleneck. A second approach is to attempt to decode the feedback packet at each DBS node separately, and to assume that successful decoding at any node will enable all other nodes to obtain the feedback via broadcast on a fast local area network (LAN). The third, and most complex, is distributed receive beamforming. Digitization and local transmission of the received signals at the
N DBS nodes to a centralized processor requires that the LAN speed scale with
N. It has been shown in [
34] that much of the received beamforming gains (within 2 dB of ideal) can be obtained even if hard decisions are exchanged: this still requires LAN speed scaling with
N, but at a smaller rate. Amplify-forward approaches for receive beamforming which sidestep such local communication by enabling on-air combining have also been proposed and demonstrated [
4]. We therefore consider ideal receive beamforming as providing a performance benchmark for the feedback channel that may be attainable with sufficient engineering effort.
The bandwidth on the feedback link may be different (typically smaller, since multiple user nodes may be sending feedback to the DBS) from that on the downlink. We average the spectral efficiency across this bandwidth when determining outage rates.
In the first approach, for a discrete set of
subcarriers, the spectral efficiency at a given DBS node, say node
k, can be calculated as
where
is the uplink channel on the
ith subcarrier for the
kth DBS node. We can now define the
-outage rate
as usual
For the second approach, outage occurs if all of the DBS nodes are unable to decode the feedback packet:
Assuming that the channel realizations for the different nodes are i.i.d., we infer that the random variables
are i.i.d., so that
. Thus, we obtain that the outage rate satisfies
which allows the individual outage probability at any DBS node to be much higher. We have checked via simulations that (
18) and (
19) yield the same results for our channels, which are obtained by independent draws from the EPA model.
For the third approach (receive beamforming), the spectral efficiency is given by
and the outage rate satisfies
Figure 11 shows the outage rates for the three approaches computed over a
MHz bandwidth for the feedback link. For a per-node SNR of
dB, the outage rates are given by
,
, and
bps/Hz. These translate to feedback data rates
of 132 Kbps, 0.8 Mbps, and 3 Mbps, respectively. Assuming that the feedback link is the bottleneck, and that
, we compute the corresponding frame lengths as
The minimum channel coherence times that can be supported using the three approaches are given by 90, 15, and 3.9 ms respectively.
We conclude that, even though the SNR on the feedback link is so low, the first approach is adequate for quasi-static links typical of rural broadband. However, if more sophisticated strategies such as the second or third approaches are employed, the DBS concept can be used to support moderate mobility.
7. Conclusions
The distributed base station concept presented in this paper is a promising approach for providing broadband access to remote areas, combining the benefits of massive MIMO with the superior propagation characteristics of large carrier wavelengths (e.g., white space frequencies). While distributed transmit beamforming is an intuitively plausible approach for range extension, we have identified and addressed key technical challenges in realizing this potential. First, we observe that the DBS system is inherently a low-SNR configuration, so that noise-resilient algorithms are required to realize the gains of distributed beamforming for this application. We demonstrate that algorithms that utilize deterministic or pseudo-random beamforming sequences over a designated training phase greatly outperform stochastic ascent algorithms such as the 1-bit feedback scheme and its variants. Among the deterministic algorithms, the proposed distributed orthogonal sequence training (DOST) scheme outperforms per-node training due to its capability of aggregating the training measurements of the entire array and boosting SNR through noise averaging. The DOST algorithm, by bootstrapping from power-pooling gain, scales indefinitely with the size of the distributed array, at the expense of linear growth in training overhead. Second, we have shown that narrowband beamforming strategies extend naturally to the wideband regime by training on a subset of pilot subcarriers, and then interpolating across subcarriers. Third, we observe that the feedback link can become a bottleneck if distributed receive beamforming is not employed and show that this impacts the rate of channel time variations that the distributed array can track. However, we show that it is possible to support relatively slowly varying links (e.g., with coherence times of the order of 100 ms) even in this setting and that faster channel time variations can be supported with more sophisticated uplink reception strategies.
The promising results in this paper motivate efforts to prototype and experimentally demonstrate range extension with the DBS concept. The identification of the feedback link as the bottleneck motivates serious investigation of more sophisticated distributed reception strategies in the low per-node SNR regime.