Abstract

Unmanned aerial vehicles (UAVs) are capable of serving as aerial base stations (BSs) for providing both cost-effective and on-demand wireless communications. This article investigates dynamic resource allocation of multiple UAVs enabled communication networks with the goal of maximizing long-term rewards. More particularly, each UAV communicates with a ground user by automatically selecting its communicating users, power levels and subchannels without any information exchange among UAVs. To model the uncertainty of environments, we formulate the long-term resource allocation problem as a stochastic game for maximizing the expected rewards, where each UAV becomes a learning agent and each resource allocation solution corresponds to an action taken by the UAVs. Afterwards, we develop a multi-agent reinforcement learning (MARL) framework that each agent discovers its best strategy according to its local observations using learning. More specifically, we propose an agent-independent method, for which all agents conduct a decision algorithm independently but share a common structure based on Q-learning. Finally, simulation results reveal that: 1) appropriate parameters for exploitation and exploration are capable of enhancing the performance of the proposed MARL based resource allocation algorithm; 2) the proposed MARL algorithm provides acceptable performance compared to the case with complete information exchanges among UAVs. By doing so, it strikes a good tradeoff between performance gains and information exchange overheads.

Highlights

  • A ERIAL communication networks, encouraging new innovative functions to deploy wireless infrastructure, have recently attracted increasing interests for providing high network capacity and enhancing coverage [2], [3]

  • We consider the multi-Unmanned aerial vehicles (UAVs) network deployed in a disc area with a radius rd = 500 m, where the ground users are randomly and uniformly distributed inside the disk and all UAVs are assumed to fly at a fixed altitude H = 100 m [2], [16]

  • We investigated the real-time designs of resource allocation for multi-UAV downlink networks to maximize the long-term rewards

Read more

Summary

Introduction

A ERIAL communication networks, encouraging new innovative functions to deploy wireless infrastructure, have recently attracted increasing interests for providing high network capacity and enhancing coverage [2], [3]. Unmanned aerial vehicles (UAVs), known as remotely piloted aircraft. Manuscript received September 20, 2018; revised May 17, 2019; accepted August 7, 2019. Date of publication August 20, 2019; date of current version February 11, 2020. This article was presented in part at the IEEE Proceedings of International Conference on Communication Workshops (ICCW), 2019 [1]. The associate editor coordinating the review of this article and approving it for publication was A.

Objectives
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.