Abstract

This paper investigates an unmanned aerial vehicle (UAV)-enabled maritime secure communication network, where the UAV aims to provide the communication service to a legitimate mobile vessel in the presence of multiple eavesdroppers. In this maritime communication networks (MCNs), it is challenging for the UAV to determine its trajectory on the ocean, since it cannot land or replenish energy on the sea surface, the trajectory should be pre-designed before the UAV takes off. Furthermore, the take-off location of the UAV and the sea lane of the vessel may be random, which leads to a highly dynamic environment. To address these issues, we propose two reinforcement learning schemes, Q-learning and deep deterministic policy gradient (DDPG) algorithms, to solve the discrete and continuous UAV trajectory design problem, respectively. Simulation results are provided to validate the effectiveness and superior performance of the proposed reinforcement learning schemes versus the existing schemes in the literature. Additionally, the proposed DDPG algorithm converges faster and achieves higher utilities for the UAV, compared to the Q-learning algorithm.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call