ContextAVO: Local context guided and refining poses for deep visual odometry

Rujun Song, Ran Zhu, Zhuoling Xiao*, Bo Yan

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

9 Citations (Scopus)

Abstract

Learning-based monocular visual odometry (VO) has lately drawn significant attention for its robustness to camera parameters and environmental variations. The correlation of ego-motion in the local time dimension, denoted as the local context, is crucial for alleviating accumulated errors of VO problems. Unlike most current learning-based methods, our approach, called ContextAVO, focuses on the effectiveness of local contexts to improve the estimation recovered from consecutive multiple optical flow snippets. To retain the pose consistency in the temporal domain, we design the Context-Attention Refining component to adaptively ameliorate current inference by exploiting the continuity of camera motions and aligning corresponding observations with local contexts. Besides, we employ the multi-length window to make ContextAVO more suitable for general scenarios and less dependent on the fixed length of the input snippet. Extensive experiments on outdoor KITTI, Malaga, ApolloScape, and indoor TUM RGB-D datasets have demonstrated that our approach efficiently produces competitive results against classic algorithms. It outperforms state-of-the-art methods by large margins, improving up to 7.40% and 48.56% for translational and rotational estimation, respectively.

Original languageEnglish
Pages (from-to)86-103
Number of pages18
JournalNeurocomputing
Volume533
DOIs
Publication statusPublished - 7 May 2023
Externally publishedYes

Keywords

  • Attention mechanism
  • Local context
  • Neural network
  • Pose consistency
  • Visual odometry

Fingerprint

Dive into the research topics of 'ContextAVO: Local context guided and refining poses for deep visual odometry'. Together they form a unique fingerprint.

Cite this