Abstract

The advent of 5G telecommunication systems has increased the need for stringent Quality of Service (QoS) requirements. A key resource in meeting these requirements are routers that perform efficient queue management, congestion control and flow prioritizations for effective network slicing. Configuring routers has traditionally been an expert driven process with static or rule-based configurations for individual flows. However, in dynamically varying traffic conditions, as are proposed in 5G use cases, these traditional approaches can generate sub-optimal configurations. In this paper, we propose a solution to this problem based on model-based Reinforcement Learning (RL) where the environment is modeled as a Partial Order Markov Decision Process (POMDP). The system is trained over different configurations for router ingress queue traffic policing, egress queue traffic shaping and traffic conditions to learn the state transition and observation probabilities of the POMDP model. The optimal policy of the learned model can then be deployed to generate optimal queue configurations automatically, adapting to current traffic conditions. These aspects are demonstrated over a real Ericsson use-case with configured routers for 5G slicing.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.