Share on Facebook Tweet on Twitter Share on LinkedIn Share by email
Visual Odometry and Map Correlation

Anat Levin and Richard Szeliski


In this paper, we study how estimates of ego-motion based on feature tracking (visual odometry) can be improved using a rough (low accuracy) map of where the observer has been. We call the process of aligning the visual ego-motion with the map locations as map correlation. Since absolute estimates of camera position are unreliable, we use stable local information such as change in orientation to perform the alignment. We also detect when the observer’s path has crossed back on itself, which helps improve both the visual odometry estimates and the alignment between the video and map sequences. The final alignment is computed using a graphical model whose MAP estimate is inferred using loopy belief propagation. Results are presented on a number of indoor and outdoor sequences.


Publication typeInproceedings
Published inIEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'2004)
AddressWashington, DC
PublisherIEEE Computer Society
> Publications > Visual Odometry and Map Correlation