Evaluation of a Stereo Visual Odometry Algorithm for Passenger Vehicle Navigation

2017-01-0046

03/28/2017

Features
Event
WCX™ 17: SAE World Congress Experience
Authors Abstract
Content
To reliably implement driver-assist features and ultimately self-driving cars, autonomous driving systems will likely rely on a variety of sensor types including GPS, RADAR, LASER range finders, and cameras. Cameras are an essential sensory component because they lend themselves to the task of identifying object types that a self-driving vehicle is likely to encounter such as pedestrians, cyclists, animals, other cars, or objects on the road. In this paper, we present a feature-based visual odometry algorithm based on a stereo-camera to perform localization relative to the surrounding environment for purposes of navigation and hazard avoidance. Using a stereo-camera enhances the accuracy with respect to monocular visual odometry. The algorithm relies on tracking a local map consisting of sparse 3D map points. By tracking this map across frames, the algorithm makes use of the full history of detected features which reduces the drift in the estimated motion trajectory. This is unlike traditional visual odometry algorithms that track features and thus estimate motion only from frame-to-frame. The visual odometry algorithm is evaluated using the widely used KITTI dataset and benchmarking suite (project by Karlsruhe Institute of Technology and Toyota Technological Institute). Relevant performance metrics are discussed and computed motion trajectories are shown.
Meta TagsDetails
DOI
https://doi.org/10.4271/2017-01-0046
Pages
8
Citation
Aladem, M., Rawashdeh, S., and Rawashdeh, N., "Evaluation of a Stereo Visual Odometry Algorithm for Passenger Vehicle Navigation," SAE Technical Paper 2017-01-0046, 2017, https://doi.org/10.4271/2017-01-0046.
Additional Details
Publisher
Published
Mar 28, 2017
Product Code
2017-01-0046
Content Type
Technical Paper
Language
English