TY - JOUR
T1 - GMMLoc: Structure Consistent Visual Localization with Gaussian Mixture Models
AU - Huang, Huaiyang
AU - Ye, Haoyang
AU - Sun, Yuxiang
AU - Liu, Ming
N1 - Funding Information:
Manuscript received February 24, 2020; accepted June 14, 2020. Date of publication June 25, 2020; date of current version July 11, 2020. This letter was recommended for publication by Associate Editor H. Ryu and Editor Y. Choi upon evaluation of the reviewers’ comments. This work was supported in part by the National Natural Science Foundation of China under Project U1713211 and in part by the Research Grant Council of Hong Kong under Project 11210017. (Corresponding author: Ming Liu.) The authors are with RAM-LAB, the Hong Kong University of Science and Technology, Hong Kong 999077, China (e-mail: [email protected]; [email protected]; [email protected]; [email protected]).
Publisher Copyright:
© 2016 IEEE.
PY - 2020/10
Y1 - 2020/10
N2 - Incorporating prior structure information into the visual state estimation could generally improve the localization performance. In this letter, we aim to address the paradox between accuracy and efficiency in coupling visual factors with structure constraints. To this end, we present a cross-modality method that tracks a camera in a prior map modelled by the Gaussian Mixture Model (GMM). With the pose estimated by the front-end initially, the local visual observations and map components are associated efficiently, and the visual structure from the triangulation is refined simultaneously. By introducing the hybrid structure factors into the joint optimization, the camera poses are bundle-adjusted with the local visual structure. By evaluating our complete system, namely GMMLoc, on the public dataset, we show how our system can provide a centimeter-level localization accuracy with only trivial computational overhead. In addition, the comparative studies with the state-of-the-art vision-dominant state estimators demonstrate the competitive performance of our method.
AB - Incorporating prior structure information into the visual state estimation could generally improve the localization performance. In this letter, we aim to address the paradox between accuracy and efficiency in coupling visual factors with structure constraints. To this end, we present a cross-modality method that tracks a camera in a prior map modelled by the Gaussian Mixture Model (GMM). With the pose estimated by the front-end initially, the local visual observations and map components are associated efficiently, and the visual structure from the triangulation is refined simultaneously. By introducing the hybrid structure factors into the joint optimization, the camera poses are bundle-adjusted with the local visual structure. By evaluating our complete system, namely GMMLoc, on the public dataset, we show how our system can provide a centimeter-level localization accuracy with only trivial computational overhead. In addition, the comparative studies with the state-of-the-art vision-dominant state estimators demonstrate the competitive performance of our method.
KW - Localization
KW - SLAM
KW - visual-based navigation
UR - http://www.scopus.com/inward/record.url?scp=85088503594&partnerID=8YFLogxK
U2 - 10.1109/LRA.2020.3005130
DO - 10.1109/LRA.2020.3005130
M3 - Journal article
AN - SCOPUS:85088503594
SN - 2377-3766
VL - 5
SP - 5043
EP - 5050
JO - IEEE Robotics and Automation Letters
JF - IEEE Robotics and Automation Letters
IS - 4
M1 - 9126150
ER -