Abstract

Imitation learning for the end-to-end autonomous driving has drawn renewed attention from academic communities. Current methods either only use images as the input, which will yield ambiguities when a vehicle approaches an intersection, or use additional command information to navigate the vehicle but inefficiently. Focusing on making the vehicle automatically drive along the given path, we propose a new and effective navigation command called as subgoal angle which does not require human participation and is calculated by the current position and subgoal of the ego-vehicle. Thus, the subgoal angle contains more information than the navigation command represented as a one-hot vector. Additionally, we propose a model architecture called as angle branched network that makes predictions based on the subgoal angle. In this network, the subgoal angle is not only used for extracting useful features but also for guiding the appropriate prediction layer to make predictions for both the steer angle and the throttle status (which controls the acceleration). Experiments are conducted in a three-dimensional urban simulator. Both quantitive and qualitive results show the effectiveness of the navigation command and the angle branched network. Moreover, the performance can be further boosted by integrating both semantic and depth information into the driving model. Especially by using the depth information, collisions with vehicles and pedestrians can be reduced.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.