Abstract
Deploying distributed cache in cognitive radio networks (CRNs), which spreads popular contents to the edge of network during the off-peak time through spectrum sharing, can reduce the deliver delay to users nearby without causing severe interference to the primary network. However, due to the un-predicable contents requirement as well as the band occupation of primary users, it is non-trivial to optimize the cache storage and contents fetching strategy of users dynamically. The letter proposes a heterogeneous multi-agent deep deterministic policy gradient (MADDPG) approach, which takes users and cache servers as two different types of agents to learn the cooperation and competition for mutual benefits. The numeral simulation demonstrates that comparing with the other single or homogeneous deep reinforcement learning (DRL) approaches, the proposed heterogeneous MADDPG can further reduce the delivery delay of users and enhance the cache efficiency of SBSs.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.