Abstract

It is still a challenge to realize safe and fast autonomous driving through deep reinforcement learning. Most autonomous driving reinforcement learning models are subject to a single experience replay approach for training agents and how to improve the driving speed and safety of agent has become the focus of research. Therefore, we present an improved Double Bias Experience Replay (DBER) approach, which enables the agent to choose its own driving learning tendency. A new loss function is proposed to ameliorate the relationship between negative loss and positive loss. The proposed approach has been applied to three algorithms to verify: Deep Q Network (DQN), Dueling Double DQN (DD-DQN) and Quantile Regression DQN (QR-DQN). Compared with the existing approaches, the proposed approach show better performance and robustness of driving policy on the driving simulator, which is implemented by the Unity ML-agents. The approach makes the vehicle agent obtain better performance, such as higher reward, faster driving speed, less lane changing and more in the same training time.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call