Abstract

With the rapid development of vehicular networks, there is an increasing demand for extensive networking, computting, and caching resources. How to allocate multiple resources effectively and efficiently for dynamic vehicular networks is extremely important. Most existing works on resource management in vehicular networks assume static network conditions. In this paper, we propose a general framework that can enable fast-adaptive resource allocation for dynamic vehicular environments. Specifically, we model the dynamics of the vehicular environment as a series of related Markov Decision Processes (MDPs), and we combine hierarchical reinforcement learning with meta-learning, which makes our proposed framework quickly adapt to a new environment by only fine-tuning the top-level master network, and meanwhile the low-level sub-networks can make the right resource allocation policy. Extensive simulation results show the effectiveness of our proposed framework, which can quickly adapt to different scenarios, and significantly improve the performance of resource management in dynamic vehicular networks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call