Abstract

The traffic scene understanding is the core technology in Intelligent Transportation Systems (ITS) and Advanced Driver Assistance System (ADAS), and it is becoming increasingly important for smart or autonomous vehicles. The recent methods for traffic scene understanding, such as Traffic Sign Recognition (TSR), Pedestrian Detection, and Vehicle Detection, have three major shortcomings. First, most models are customized for recognizing a specific category of traffic target instead of general traffic targets. Second, as for these recognition modules, the task of traffic scene understanding is to recognize objects rather than make driving suggestions or strategies. Third, numerous independent recognition modules disadvantage to fusing multi-modal information to make a comprehensive decision for driving operation in accordance with complicated traffic scenes. In this paper, we first introduce the image captioning model to alleviate the aforementioned shortcomings. Different from existing methods, our primary idea is to accurately identify all categories of traffic objects and understand traffic scenes by making full use of all information, and making the suggestions or strategy for driving operation in natural language by using Long Short Term Memory network (LSTM) rather than keywords. The proposed solution naturally solves the problems of feature fusion, general object recognition, and low-level semantic understanding. We tested the solution on our created traffic scene image dataset for evaluation of image captioning. Extensive experiments including quantitative and qualitative comparisons demonstrate that the proposed solution can identify more objects and produce higher-level semantic information than the state-of-the-arts.

Highlights

  • Developing new Intelligent Transportation Systems (ITS) which take into consideration the socio-economical, environmental, and safety factors of the modern society is one of the grand challenges of the century

  • We propose a solution for traffic scene understanding with the image captioning method

  • Our goal is to describe traffic scenes using image captioning based on the above framework, so we construct the architecture of our model

Read more

Summary

Introduction

Developing new Intelligent Transportation Systems (ITS) which take into consideration the socio-economical, environmental, and safety factors of the modern society is one of the grand challenges of the century. Autonomous vehicles are promising for improving the safety and efficiency of transportation and mobility systems by providing vehicle control during normal driving. A human driver will still be expected to have control responsibilities for an automated driving car in the case of an emergency. The most possible solution to ITS may be mixed autonomous vehicles and human-driven vehicles rather than fully automated without human operation. The Advanced Driver Assistance System (ADAS) plays an important role in the Intelligent Transportation System.

Objectives
Methods
Findings
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.