Caching popular content at the edge of wireless networks leads to backhaul congestion mitigation. To come up with an effective caching policy, content popularity distribution should be taken into account, which is not accurately known in most practical scenarios. Moreover, the mobile users’ (MU) request pattern may not always follow a well-defined distribution since some malicious MUs may deliberately issue their requests incompatible with the content popularity statistics. In this paper, we consider the problem of cache content placement in a 5G mmWave small cell network that relies on integrated access and backhaul (IAB) technology for pushing contents to MUs. We assume that the IAB node is equipped with a cache and has no prior knowledge about the content popularity profiles; instead, it only relies on the observation of the instantaneous demands to shape its caching policy. Also, malicious MUs may exist whose goals are to increase cache miss by issuing fictitious requests. The IAB node decides on which contents to cache and for how long, given that frequently replacing contents incurs administrative costs. We model the content placement problem as an ”adversarial combinatorial multi-armed bandit process with switching costs (ACMAB-SC)” and present an online learning algorithm for shaping the caching policy. We conduct extensive simulation experiments to evaluate the convergence property and assess the performance of our algorithm in terms of backhaul congestion, delay, and cache hit ratio. We also compare against two baseline online learning schemes, including a CMAB-based approach and a generic caching policy based on the ”Follow the Perturbed Leader (FPL)” algorithm.
Read full abstract