Abstract

Building a human-like car-following model that can accurately simulate drivers’ car-following behaviors is helpful to the development of driving assistance systems and autonomous driving. Recent studies have shown the advantages of applying reinforcement learning methods in car-following modeling. However, a problem has remained where it is difficult to manually determine the reward function. This paper proposes a novel car-following model based on generative adversarial imitation learning. The proposed model can learn the strategy from drivers’ demonstrations without specifying the reward. Gated recurrent units was incorporated in the actor-critic network to enable the model to use historical information. Drivers’ car-following data collected by a test vehicle equipped with a millimeter-wave radar and controller area network acquisition card was used. The participants were divided into two driving styles by K-means with time-headway and time-headway when braking used as input features. Adopting five-fold cross-validation for model evaluation, the results show that the proposed model can reproduce drivers’ car-following trajectories and driving styles more accurately than the intelligent driver model and the recurrent neural network-based model, with the lowest average spacing error (19.40%) and speed validation error (5.57%), as well as the lowest Kullback-Leibler divergences of the two indicators used for driving style clustering.

Highlights

  • With the process of urbanization and the rapid growth in the number of vehicles, car following has become the most common driving behavior in daily driving

  • The structure of the proposed model denoted as generative adversarial imitation learning (GAIL)-gated recurrent units (GRU) [16,39] is shown in Figure 1 which is composed of two parts: (1) a generator that consists of an actor-critic structure and the car-following environment, which generates the faked state-action pairs, and (2) a discriminator that functions as classifying the generated state-action pairs and the actual state-action pairs obtained from the collected data

  • The performances of the proposed model GAIL-GRU, the theoretical-driven model IDM, and the recent behavior-cloning model recurrent neural network (RNN) for replicating the car-following trajectories of the drivers with different driving styles on the training sets and test sets are presented in Figures 12 and 13, respectively

Read more

Summary

Introduction

With the process of urbanization and the rapid growth in the number of vehicles, car following has become the most common driving behavior in daily driving. The error existing in the single time-step prediction will accumulate and gradually increase in the sequential decision-making process, which may cause the model to reach unseen states, making the model have even worse predictions To avoid this problem, some researchers have begun to use reinforcement learning (RL) methods [11,14]. A recently proposed algorithm called generative adversarial imitation learning (GAIL) was applied to model drivers’ car-following strategies. The proposed model uses a nonlinear function that uses neural networks to automatically learn drivers’ rewards and strategies, and the training of the model does not need to solve the RL subprocess, which can save a lot of computation power.

Theoretical-Driven Car-Following Models
Behavior Cloning Car-Following Models
Reinforcement Learning
Inverse Reinforcement Learning
Generative Adversarial Imitation Learning
The Proposed Model
The Generator
The Discriminator
The Proposed Algorithm
2: Algorithm begins: 3
The Experiments
Car-Following Periods Extraction
Driving Style Clustering Based on K-Means
Simulation Setup
Root Mean Square Percentage Error
Kullback-Leibler Divergence
Cross-Validation
Models Investigated
RNN Based Model
Results
Discussion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call