Abstract

Backgrounds: The traffic signal control (TSC) system could be more intelligently controlled by deep reinforcement learning (DRL) and information provided by connected and automated vehicles (CAVs). However, the direct training procedure of the DRL is time-consuming and hard to converge. Methods: This study improves the training efficiency of the deep Q network (DQN) by transferring the well-trained action policy of a previous DQN model into a target model under similar traffic scenarios. Different reward parameters, exploration rates, and action step lengths are tested. The performance of the transfer-based DQN-TSC is analyzed by considering different traffic demands and market penetration rates (MPRs) of CAVs. The information level requirements of the DQN-TSC are also investigated. Results: Compared to directly trained DQN, transfer-based models could improve both the training efficiency and model performance. In high traffic scenarios with a 100% MPR of CAVs, the total waiting time, CO2 emission, and fuel consumption in the transfer-based TSC decrease about 38%, 34%, and 34% compared to pre-timed signal schemes. Also, the transfer-based TSC system requires more than 20% to 40% MPRs of CAVs under different traffic demands to perform better than pre-timed signal schemes. Conclusions: The proposed model could improve both the traffic performance of the TSC system and the training efficiency of the DQN model. The insights of this study should be helpful to planners and engineers in designing intelligent signal intersections and providing guidance for engineering applications of the DQN TSC systems.

Highlights

  • With the rapid development of learning-based artificial intelligence technologies, combining the management of transportation systems with reinforcement learning (RL) technologies provides a new potential solution to improve the efficiency, safety, and sustainability of intelligent transportation systems

  • COMPARISON BETWEEN DIRECT AND TRANSFERBASED LEARNING To test the efficiency of the transfer-based deep Q network (DQN) approach, a comparison between direct training and transfer-based training with full exploration (ε-greedy from 1 to 0) is made under a scenario with medium traffic and a 100% market penetration rates (MPRs) of connected and autonomous vehicles (CAVs)

  • This paper tests the impacts of information levels of the mixed traffic on the transfer-based DQN traffic signal control (TSC) system

Read more

Summary

Introduction

With the rapid development of learning-based artificial intelligence technologies, combining the management of transportation systems with reinforcement learning (RL) technologies provides a new potential solution to improve the efficiency, safety, and sustainability of intelligent transportation systems. The emerging development of the vehicle to infrastructure (V2I) communication technology enables connected vehicles (CVs) or connected and autonomous vehicles (CAVs) to transmit real-time information on vehicles to the traffic signal control (TSC) system. All these technologies make it feasible to control an intelligent TSC system by RL technologies. MPRs of CAVs could determine the information level of vehicles that can be obtained by the TSC system. To investigate the validity and information-level requirement of the RL-controlled TSC system, it is important to study the impacts of different MPRs of CAVs on the RL-based TSC system

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call