Abstract

Efficient allocation of resources is crucial for optimizing wireless networks that face constraints in bandwidth, power, and spectrum. This paper proposes a decentralized reinforcement learning (RL) model that departs from traditional centralized paradigms to revolutionize resource optimization. The proposed model empowers individual wireless devices with autonomous decision-making capabilities, enhancing adaptability and scalability by leveraging Deep Q-Network (DQN) and Proximal Policy Optimization (PPO). The innovative integration of memory mechanisms facilitates learning from past experiences, addressing the dynamic nature of wireless environments. This decentralized RL model offers practical implications for improved efficiency, adaptability, and reliability in wireless network resource optimization. By transforming individual devices into collaborative decision-makers, our proposed model contributes to a resilient and responsive wireless communication infrastructure. The specific contributions of this paper include the pioneering use of DQN and PPO algorithms within a multi-agent system, offering a groundbreaking solution for dynamic resource optimization in wireless networks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call