Abstract
The goal in this work is to design online power control policies for large energy harvesting (EH) networks where, due to large energy overhead involved in the exchange of state information among the nodes, it is infeasible to use a centralized policy. Furthermore, typical applications of EH networks concern the scenario where the statistical information, about both the EH process and the wireless channel, is not available. In order to address these challenges, we propose a mean-field multiagent deep reinforcement learning framework. The proposed approach enables the nodes to learn online power control policies in a fully distributed fashion, i.e., it does not require the nodes to exchange the information about their states. Using the underlying structure of the problem, we analytically establish the convergence of the proposed scheme. In particular, we show that the policies obtained using the proposed approach converge to the `stationary' Nash equilibrium. Our simulation results illustrate the efficacy of the power control policies, learned through the proposed approach. In particular, the mean-field multi-agent reinforcement learning scheme achieves a performance close to the state-of-the-art centralized policies which operate using the information about the state of whole network.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.