Title:
Visual Odometry Priors for robust EKF-SLAM

dc.contributor.author Alcantarilla, Pablo F.
dc.contributor.author Bergasa, Luis Miguel
dc.contributor.author Dellaert, Frank
dc.contributor.corporatename Georgia Institute of Technology. Center for Robotics and Intelligent Machines
dc.contributor.corporatename Georgia Institute of Technology. College of Computing
dc.contributor.corporatename Georgia Institute of Technology. School of Interactive Computing
dc.contributor.corporatename Universidad de Alcalá. Departamento de Electrónica
dc.date.accessioned 2011-03-28T22:04:04Z
dc.date.available 2011-03-28T22:04:04Z
dc.date.issued 2010
dc.description ©2010 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works. en_US
dc.description Presented at the 2010 IEEE International Conference on Robotics and Automation (ICRA), 3-7 May 2010, Anchorage, AK.
dc.description DOI: 10.1109/ROBOT.2010.5509272
dc.description.abstract One of the main drawbacks of standard visual EKF-SLAM techniques is the assumption of a general camera motion model. Usually this motion model has been implemented in the literature as a constant linear and angular velocity model. Because of this, most approaches cannot deal with sudden camera movements, causing them to lose accurate camera pose and leading to a corrupted 3D scene map. In this work we propose increasing the robustness of EKF-SLAM techniques by replacing this general motion model with a visual odometry prior, which provides a real-time relative pose prior by tracking many hundreds of features from frame to frame. We perform fast pose estimation using the two-stage RANSAC-based approach from [1]: a two-point algorithm for rotation followed by a one-point algorithm for translation. Then we integrate the estimated relative pose into the prediction step of the EKF. In the measurement update step, we only incorporate a much smaller number of landmarks into the 3D map to maintain real-time operation. Incorporating the visual odometry prior in the EKF process yields better and more robust localization and mapping results when compared to the constant linear and angular velocity model case. Our experimental results, using a stereo camera carried in hand as the only sensor, clearly show the benefits of our method against the standard constant velocity model. en_US
dc.identifier.citation Alcantarilla, P.F., Bergasa, L.M., & Dellaert, F. (2010). “Visual Odometry Priors for Robust EKF-SLAM". Proceedings of the IEEE International Conference on Robotics and Automation (ICRA 2010), 3-7 May 2010, 3501-3506. en_US
dc.identifier.issn 1050-4729
dc.identifier.uri http://hdl.handle.net/1853/38308
dc.language.iso en_US en_US
dc.publisher Georgia Institute of Technology en_US
dc.publisher Institute of Electrical and Electronics Engineers
dc.subject Extended Kalman Filter (EKF) en_US
dc.subject Pose estimation en_US
dc.subject Sensors en_US
dc.subject Tracking en_US
dc.title Visual Odometry Priors for robust EKF-SLAM en_US
dc.type Text
dc.type.genre Post-print
dc.type.genre Proceedings
dspace.entity.type Publication
local.contributor.author Dellaert, Frank
local.contributor.corporatename Institute for Robotics and Intelligent Machines (IRIM)
relation.isAuthorOfPublication dac80074-d9d8-4358-b6eb-397d95bdc868
relation.isOrgUnitOfPublication 66259949-abfd-45c2-9dcc-5a6f2c013bcf
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
Alcantarilla10icra1.pdf
Size:
769.67 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.76 KB
Format:
Item-specific license agreed upon to submission
Description: