Abstract
Many intelligent transportation system (ITS) applications will increasingly rely on lane-level vehicle positioning that requires high accuracy, bandwidth, availability, and integrity. Lane-level positioning methods must reliably work in real time in a wide range of environments, spanning rural to urban areas. Traditional positioning sensors such as the Global Navigation Satellite Systems may have poor performance in dense urban areas, where obstacles block satellite signals. This paper presents a sensor fusion technique that uses computer vision and differential pseudorange Global Positioning System (DGPS) measurements to aid an inertial navigation system (INS) in challenging environments where GPS signals are limited and/or unreliable. To supplement limited DGPS measurements, this method uses mapped landmarks that were measured through a priori observations (e.g., traffic light location data), taking advantage of existing infrastructure that is abundant within suburban/urban environments. For example, traffic lights are easily detected by color vision sensors in both day and night conditions. A tightly coupled estimation process is employed to use observables from satellite signals and known feature observables from a camera to correct an INS that is formulated as an extended Kalman filter. A traffic light detection method is also outlined, where the projected feature uncertainty ellipse is utilized to perform data association between a predicted feature and a set of detected features. Real-time experimental results from real-world settings are presented to validate the proposed localization method.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: IEEE Transactions on Intelligent Transportation Systems
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.