Abstract
Unmanned aerial vehicles (UAVs) are capable of serving as aerial base stations (BSs) for providing both cost-effective and on-demand wireless communications. This article investigates dynamic resource allocation of multiple UAVs enabled communication networks with the goal of maximizing long-term rewards. More particularly, each UAV communicates with a ground user by automatically selecting its communicating users, power levels and subchannels without any information exchange among UAVs. To model the uncertainty of environments, we formulate the long-term resource allocation problem as a stochastic game for maximizing the expected rewards, where each UAV becomes a learning agent and each resource allocation solution corresponds to an action taken by the UAVs. Afterwards, we develop a multi-agent reinforcement learning (MARL) framework that each agent discovers its best strategy according to its local observations using learning. More specifically, we propose an agent-independent method, for which all agents conduct a decision algorithm independently but share a common structure based on Q-learning. Finally, simulation results reveal that: 1) appropriate parameters for exploitation and exploration are capable of enhancing the performance of the proposed MARL based resource allocation algorithm; 2) the proposed MARL algorithm provides acceptable performance compared to the case with complete information exchanges among UAVs. By doing so, it strikes a good tradeoff between performance gains and information exchange overheads.
Highlights
A ERIAL communication networks, encouraging new innovative functions to deploy wireless infrastructure, have recently attracted increasing interests for providing high network capacity and enhancing coverage [2], [3]
We consider the multi-Unmanned aerial vehicles (UAVs) network deployed in a disc area with a radius rd = 500 m, where the ground users are randomly and uniformly distributed inside the disk and all UAVs are assumed to fly at a fixed altitude H = 100 m [2], [16]
We investigated the real-time designs of resource allocation for multi-UAV downlink networks to maximize the long-term rewards
Summary
A ERIAL communication networks, encouraging new innovative functions to deploy wireless infrastructure, have recently attracted increasing interests for providing high network capacity and enhancing coverage [2], [3]. Unmanned aerial vehicles (UAVs), known as remotely piloted aircraft. Manuscript received September 20, 2018; revised May 17, 2019; accepted August 7, 2019. Date of publication August 20, 2019; date of current version February 11, 2020. This article was presented in part at the IEEE Proceedings of International Conference on Communication Workshops (ICCW), 2019 [1]. The associate editor coordinating the review of this article and approving it for publication was A.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.