Abstract

Adaptive packet scheduling can efficiently enhance the performance of multipath Data Transmission. However, realizing precise packet scheduling is challenging due to the nature of high dynamics and unpredictability of network link states. To this end, this paper proposes a distributed asynchronous deep reinforcement learning framework to intensify the dynamics and prediction of adaptive packet scheduling. Our framework contains two parts: local asynchronous packet scheduling and distributed cooperative control center. In local asynchronous packet scheduling, an asynchronous prioritized replay double deep Q-learning packets scheduling algorithm is proposed for dynamic adaptive packet scheduling learning, which makes a combination of prioritized replay double deep Q-learning network (P-DDQN) to make the fitting analysis. In distributed cooperative control center, a distributed scheduling learning and neural fitting acceleration algorithm to adaptively update neural network parameters of P-DDQN for more precise packet scheduling. Experimental results show that our solution has a better performance than Random weight algorithm and Round-Robin algorithm in throughput and loss ratio. Further, our solution has 1.32 times and 1.54 times better than Random weight algorithm and Round-Robin algorithm on the stability of multipath data transmission, respectively.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call