Abstract

In this thesis, we propose an environment perception framework for autonomous driving using deep reinforcement learning (DRL) that exhibits learning in autonomous vehicles under complex interactions with the environment, without being explicitly trained on driving datasets. Unlike existing techniques, our proposed technique takes the learning loss into account under deterministic as well as stochastic policy gradient. We apply DRL to object detection and safe navigation while enhancing a self-driving vehicle’s ability to discern meaningful information from surrounding data. For efficient environmental perception and object detection, various Q-learning based methods have been proposed in the literature. Unlike other works, this thesis proposes a collaborative deterministic as well as stochastic policy gradient based on DRL. Our technique is a combination of variational autoencoder (VAE), deep deterministic policy gradient (DDPG), and soft actor-critic (SAC) that adequately trains a self-driving vehicle. In this work, we focus on uninterrupted and reasonably safe autonomous driving without colliding with an obstacle or steering off the track. We propose a collaborative framework that utilizes best features of VAE, DDPG, and SAC and models autonomous driving as partly stochastic and partly deterministic policy gradient problem in continuous action space, and continuous state space. To ensure that the vehicle traverses the road over a considerable period of time, we employ a reward-penalty based system where a higher negative penalty is associated with an unfavourable action and a comparatively lower positive reward is awarded for favourable actions. We also examine the variations in policy loss, value loss, reward function, and cumulative reward for ‘VAE+DDPG’ and ‘VAE+SAC’ over the learning process.

Highlights

  • Autonomous driving (AD) has received significant attention from the research community as well as the industry, and it forms a critical component of the connected and autonomous vehicles (CAV) and internet of vehicles (IoV) framework [1], [2]

  • Reinforcement learning combined with deep learning, known as deep reinforcement learning (DRL) has emerged as a strong artificial intelligence paradigm that has been successfully tested on various gaming applications [7]

  • The following section presents the performance analysis for variational auto encoder (VAE)+deep deterministic policy gradient (DDPG) and VAE+soft actor-critic (SAC) schemes followed by summary of the findings

Read more

Summary

Introduction

Autonomous driving (AD) has received significant attention from the research community as well as the industry, and it forms a critical component of the connected and autonomous vehicles (CAV) and internet of vehicles (IoV) framework [1], [2]. In order to deal with the real-life driving scenarios, to safely navigate the traffic, to reduce travel delay, and to avoid congestion, novel object detection and scene perception techniques are required [2]. Environment perception and cognition to gain a comprehensive understanding of the driving scenarios are considered critical enablers for connected and autonomous vehicles. One technique to enhance perception in autonomous vehicles is reinforcement learning that teaches machines through constant interaction with the environment, learning from past experiences, and subsequently improving upon them [6]. A prevalent data-driven machine learning (ML) technique known as supervised learning allows these agents to learn from a pre-existing collection of data. The deep reinforcement learning (DRL) methodology, a combination of the existing deep learning (DL) and reinforcement learning (RL) scheme, is currently being studied as a baseline format for the self-driving vehicles [11].

Objectives
Methods
Results
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.