Abstract
In this paper we consider a large deviations problem for a discrete-time polling system consisting of two-parallel queues and a single server. The arrival process of each queue is a superposition of traffic streams generated by a number of mutually independent and identical Markovian on/off sources, and the single server serves the two queues according to the so-called Bernoulli service schedule. Using the large de- viations techniques, we derive the upper and lower bounds of the probability that the queue length of each queue exceeds a certain level (i.e., the buffer overflow probabil- ity). These results have important implications for traffic management of high-speed communication networks such as call admission control and bandwidth allocation.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.