GMMLoc: Structure Consistent Visual Localization with Gaussian Mixture Models release_iis7xjxkkrcohhwto76dw22qty

by Huaiyang Huang, Haoyang Ye, Yuxiang Sun, Ming Liu

Released as a article .



Incorporating prior structure information into the visual state estimation could generally improve the localization performance. In this letter, we aim to address the paradox between accuracy and efficiency in coupling visual factors with structure constraints. To this end, we present a cross-modality method that tracks a camera in a prior map modelled by the Gaussian Mixture Model (GMM). With the pose estimated by the front-end initially, the local visual observations and map components are associated efficiently, and the visual structure from the triangulation is refined simultaneously. By introducing the hybrid structure factors into the joint optimization, the camera poses are bundle-adjusted with the local visual structure. By evaluating our complete system, namely GMMLoc, on the public dataset, we show how our system can provide a centimeter-level localization accuracy with only trivial computational overhead. In addition, the comparative studies with the state-of-the-art vision-dominant state estimators demonstrate the competitive performance of our method.
In text/plain format

Archived Files and Locations

application/pdf  3.3 MB
file_ftsgdjl2szfrbnywh6x6w3liry (repository) (webarchive)
Read Archived PDF
Preserved and Accessible
Type  article
Stage   submitted
Date   2020-06-24
Version   v1
Language   en ?
Work Entity
access all versions, variants, and formats of this works (eg, pre-prints)
Catalog Record
Revision: c906d4e1-279a-47ae-8f49-03a2be18faef