One of the key ideas for reducing downlink channel acquisition overhead for FDD massive MIMO systems is to exploit a combination of two assumptions: (i) the dimension of channel models in propagation domain may be much smaller than the next-generation base-station array sizes (e.g., 64 or more antennas), and (ii) uplink and downlink channels may share the same low-dimensional propagation domain. Our channel measurements demonstrate that the two assumptions may not always hold, thereby impacting the predicted performance of methods that rely on the above assumptions. In this paper, we analyze the error in modeling the downlink channel using uplink measurements, caused by the mismatch from the above two assumptions. We investigate how modeling error varies with base-station array size and provide both numerical and experimental results. We observe that modeling error increases with the number of base-station antennas, and channels with larger angular spreads have larger modeling error. Utilizing our modeling error analysis, we then investigate the resulting beamforming performance rate loss. Accordingly, we observe that the rate loss increases with the number of base-station antennas, and channels with larger angular spreads suffer from higher rate loss.
Read full abstract