Abstract

The limited battery capacity and low computing capability of wireless Internet of Things (IoT) devices can hardly support computation-intensive and delay-sensitive applications. While recent development of wireless power transfer (WPT) and mobile edge computing (MEC) technologies help IoT devices harvest energy and offload computation tasks to edge servers. While it is still challenging to design an efficient offloading policy to improve the performance of the IoT network. In this article, we consider a MEC network that has WPT capability and adopts the non-orthogonal multiple access (NOMA) technology to offload tasks partially. Our goal is to propose an online algorithm to optimize resource allocation under a wireless dynamic channel scenario. In order to obtain the optimal offloading decision and resource allocation efficiently, we propose a Deep Reinforcement learning-based Online Sample-improving (DROS) framework which implements a deep neural network to input the discretized channel gains to obtain the optimal WPT duration. Based on the WPT duration derived by DNN, we design an optimization algorithm to derive the optimal energy proportion for offloading data. Numerical results verify that compared with traditional optimization algorithms, our proposed DROS has significantly sped up convergence for better solutions.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call