By Topic

Real-Time Computer Vision/DGPS-Aided Inertial Navigation System for Lane-Level Vehicle Navigation

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

5 Author(s)
Anh Vu ; Dept. of Electr. Eng., Univ. of California, Riverside, CA, USA ; Ramanandan, A. ; Anning Chen ; Farrell, J.A.
more authors

Many intelligent transportation system (ITS) applications will increasingly rely on lane-level vehicle positioning that requires high accuracy, bandwidth, availability, and integrity. Lane-level positioning methods must reliably work in real time in a wide range of environments, spanning rural to urban areas. Traditional positioning sensors such as the Global Navigation Satellite Systems may have poor performance in dense urban areas, where obstacles block satellite signals. This paper presents a sensor fusion technique that uses computer vision and differential pseudorange Global Positioning System (DGPS) measurements to aid an inertial navigation system (INS) in challenging environments where GPS signals are limited and/or unreliable. To supplement limited DGPS measurements, this method uses mapped landmarks that were measured through a priori observations (e.g., traffic light location data), taking advantage of existing infrastructure that is abundant within suburban/urban environments. For example, traffic lights are easily detected by color vision sensors in both day and night conditions. A tightly coupled estimation process is employed to use observables from satellite signals and known feature observables from a camera to correct an INS that is formulated as an extended Kalman filter. A traffic light detection method is also outlined, where the projected feature uncertainty ellipse is utilized to perform data association between a predicted feature and a set of detected features. Real-time experimental results from real-world settings are presented to validate the proposed localization method.

Published in:

Intelligent Transportation Systems, IEEE Transactions on  (Volume:13 ,  Issue: 2 )