Abstract
Connected devices generate large amount of data for IoT applicatons. Assisted by edge computing, caching IoT data at the edge nodes is considered as a promising technique for its advantage in reducing network traffic and service delay of cloud platform. However, the IoT data is characterized by transient lifetime and cache capacity that is limited by the edge nodes. As a consequence, caching policy should consider both data transiency and storage capacity of edge nodes. Inspired by the success of deep reinforcement learning (DRL) in deal with Markov Decision Process (MDP) problem in unknown environment, A DRL-based algorithm for edge caching problem is proposed in this paper. The proposed Advantage Actor Critic (A2C)-based algorithm is aimed at maximizing the long-term energy saving without knowledge of the IoT data popularity profiles. Simulation results demonstrate that the proposed DRL-based algorithm can achieve higher energy saving and cache hit ratio compared with the baseline algorithms.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.