TY - GEN
T1 - Feature-Based and Convolutional Neural Network Fusion Method for Visual Relocalization
AU - Wang, Li
AU - Li, Ruifeng
AU - Sun, Jingwen
AU - Soon Seah, Hock
AU - Quah, Chee Kwang
AU - Zhao, Lijun
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/12/18
Y1 - 2018/12/18
N2 - Relocalization is one of the necessary modules for mobile robots in long-term autonomous movement in an environment. Currently, visual relocalization algorithms mainly include feature-based methods and CNN-based (Convolutional Neural Network) methods. Feature-based methods can achieve high localization accuracy in feature-rich scenes, but the error is quite large or it even fails in cases with motion blur, texture-less scene and changing view angle. CNN-based methods usually have better robustness but poor localization accuracy. For this reason, a visual relocalization algorithm that combines the advantages of the two methods is proposed in this paper. The BoVW (Bag of Visual Words) model is used to search for the most similar image in the training dataset. PnP (Perspective n Points) and RANSAC (Random Sample Consensus) are employed to estimate an initial pose. Then the number of inliers is utilized as a criterion whether the feature-based method or the CNN-based method is to be leveraged. Compared with a previous CNN-based method, PoseNet, the average position error is reduced by 45.6% and the average orientation error is reduced by 67.4% on Microsoft's 7-Scenes datasets, which verifies the effectiveness of the proposed algorithm.
AB - Relocalization is one of the necessary modules for mobile robots in long-term autonomous movement in an environment. Currently, visual relocalization algorithms mainly include feature-based methods and CNN-based (Convolutional Neural Network) methods. Feature-based methods can achieve high localization accuracy in feature-rich scenes, but the error is quite large or it even fails in cases with motion blur, texture-less scene and changing view angle. CNN-based methods usually have better robustness but poor localization accuracy. For this reason, a visual relocalization algorithm that combines the advantages of the two methods is proposed in this paper. The BoVW (Bag of Visual Words) model is used to search for the most similar image in the training dataset. PnP (Perspective n Points) and RANSAC (Random Sample Consensus) are employed to estimate an initial pose. Then the number of inliers is utilized as a criterion whether the feature-based method or the CNN-based method is to be leveraged. Compared with a previous CNN-based method, PoseNet, the average position error is reduced by 45.6% and the average orientation error is reduced by 67.4% on Microsoft's 7-Scenes datasets, which verifies the effectiveness of the proposed algorithm.
UR - http://www.scopus.com/inward/record.url?scp=85060776600&partnerID=8YFLogxK
U2 - 10.1109/ICARCV.2018.8581204
DO - 10.1109/ICARCV.2018.8581204
M3 - Conference contribution
AN - SCOPUS:85060776600
T3 - 2018 15th International Conference on Control, Automation, Robotics and Vision, ICARCV 2018
SP - 1489
EP - 1495
BT - 2018 15th International Conference on Control, Automation, Robotics and Vision, ICARCV 2018
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 15th International Conference on Control, Automation, Robotics and Vision, ICARCV 2018
Y2 - 18 November 2018 through 21 November 2018
ER -