Abstract

In order to reduce the consumption cost for successive interference cancellation in non-orthogonal multiple access(NOMA), we propose a resource allocation scheme that involves both orthogonal multiple access and NOMA technologies. The scheme uses deep learning to choose the appropriate access according to the communication environment. Moreover, the scheme jointly allocates subcarrier and power resources for users by utilizing a deep Q network and a multi-agent deep deterministic policy gradient network. Meanwhile, an adaptive mechanism combining online learning and offline learning is introduced into allocation scheme to flexibly adapt to the communication environment. Results show that the proposed scheme can achieve better system performance in sum-rate. In order to better cope with changes in the environment and make the resource allocation strategy more robust, we propose a novel resource allocation algorithm combining transfer learning and deep reinforcement learning. The algorithm can effectively improve the model convergence speed when changing the communication environment. Furthermore, the algorithm allows us to transfer the subcarrier allocation network and the power allocation network simultaneously or separately depending on the environment.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.