Abstract

Network slicing to create multiple virtual networks, called network slice, is a promising technology to enable networking resource sharing among multiple tenants for the 5th generation (5G) networks. By offering a network slice to slice tenants, network slicing supports parallel services to meet the service level agreement (SLA). In legacy networks, every tenant pays a fixed and roughly estimated monthly or annual fee for shared resources according to a contract signed with a provider. However, such a fixed resource allocation mechanism may result in low resource utilization or violation of user quality of service (QoS) due to fluctuations in the network demand. To address this issue, we introduce a resource management system for network slicing and propose a dynamic resource adjustment algorithm based on reinforcement learning approach from each tenant’s point of view. First, the resource management for network slicing is modeled as a Markov Decision Process (MDP) with the state space, action space, and reward function. Then, we propose a Q-learning-based dynamic resource adjustment algorithm that aims at maximizing the profit of tenants while ensuring the QoS requirements of end-users. The numerical simulation results demonstrate that the proposed algorithm can significantly increase the profit of tenants compared to existing fixed resource allocation methods while satisfying the QoS requirements of end-users.

Highlights

  • Since network slices will be used by traffic engineering businesses, network slicing is a matter of business and economic model as well as a simple resource allocation mechanism

  • We propose a resource management mechanism based on variations of the traffic mix using Q-learning algorithm

  • We considered the dynamic resource trading in network slicing to maximize the profit of tenants while ensuring the quality of service (QoS) requirements of end-users

Read more

Summary

Introduction

Project (3GPP) suggests that static resource allocation based on fixed network sharing can be one of the approaches for resource management in network slicing. Such a static allocation mechanism may lead to low efficiency. We propose a resource management mechanism based on variations of the traffic mix using Q-learning algorithm. The tenant interacts with end-users using the latter interface to provide the resources to them Under such a Q-learning-based dynamic resource trading environment, each tenant exhibits a strategic behavior to maximize its own profit. We propose a Q-learning-based dynamic resource management strategy to maximize tenant’s profit while satisfying QoS of end-users in each slice from each tenant’s point of view.

Related Work
Overall System Architecture
Service and Traffic Model
Resource Trading System Using a Markov Decision Process
State Space
Action Space
Reward Function
Dynamic Resource Adjustment Algorithm
Setup of the Simulation Environment
Q-learning Algorithm Convergence
Competitiveness
Impact of the Trading Interval on the Algorithm Performance
QoS Violation WITH Flow Ratio Change
QoS Weight
Findings
Conclusions

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.