Abstract
The flexible and highly mobile unmanned aerial vehicle (UAV) with computing capabilities can improve the quality of experience (QoE) of ground users (GUs) according to real-time service requirements by performing flight maneuvers. In this study, we investigate a task offloading scheme and trajectory optimization in a multi-UAV-assisted system, where UAVs offload a portion of multiple GUs’ computational tasks. The optimization problem is formulated to jointly minimize the energy consumption of UAVs and the task latency of GUs by optimizing trajectory, task allocation and offloading proportion. This paper proposes a heterogeneous multi-agent reinforcement learning (MARL)-based approach to solve the issue in high dimensions and limited states, where UAV and GU are treated separately as two different types of agents. Due to the high cost and low sample efficiency of online training of RL algorithms, a generative adversarial network (GAN)-powered auxiliary training mechanism is proposed, which reduces the overhead of interacting with the real world and makes the agent’s policy appropriate for real-world execution environment via offline training with generated environment states. Numerical evaluation results demonstrate that the proposed algorithm outperforms other benchmark algorithms in terms of energy consumption and task latency.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.