Abstract

Integrating unmanned aerial vehicles (UAVs) with wireless systems is widely anticipated to bring enormous benefits to network users, and is considered as a key enabling technology for next generation wireless networks. However deployment of UAVs raises security concerns, as they are prone to eavesdropping attacks. In this paper we investigate secure data transmissions in a UAV-aided communication system from the physical layer security (PLS) perspective. The UAV is employed as a flying base station and transmit confidential information to a ground user. Specifically, an eavesdropper is moving in the vicinity of the ground user, attempting to intercept legitimate data transmission. We aim to optimize UAV trajectory to maximize secrecy throughput and due to the fast changing environment with unpredictable movements of the eave, this nonconvex problem is difficult to solve. Instead, we propose a deep reinforcement learning framework by reformulating this problem as a Markov Decision Process (MDP). By carefully designing state spaces, state-dependent action sets and time-dependant reward signals, we aim to help the agent learn an optimized trajectory and complete the flight task within time limit. We adapt the deep Q network (DQN) algorithm and extend it with prioritized experience replay (PER) and importance sampling method to ensure a stable and effective learning outcome. Simulation results validate the effectiveness of the learning algorithm and confirm that the agent is able to learn an effective policy with desirable and robust learning outcomes. Compared with benchmark schemes, our proposed scheme is shown to achieve considerable performance gain in secrecy throughput.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call