Abstract
Unsupervised deep learning methods have made significant progress in monocular visual odometry (VO) tasks. However, due to the complexity of the real-world scene, learning the camera ego-motion from the RGB information of monocular images in an unsupervised way is still challenging. Existing methods mainly learn motion from the original RGB information, lacking higher-level input from scene understanding. Hence, this paper proposes an unsupervised monocular VO framework that combines the instance and RGB information, named combined information based (CI-VO). The proposed method includes two stages. First is obtaining the instance maps of the monocular images, without finetuning on the VO dataset. Then we obtain the combined information from the two types of information, which is input into the proposed combined information based pose estimation network, named CI-PoseNet, to estimate the relative pose of the camera. To make better use of the two types of information, we propose a fusion feature extraction network to extract the fused features from the combined information. Experiments on the KITTI odometry and KITTI raw dataset show that the proposed method has good performance in the camera pose estimation task, which exceeds the existing mainstream methods.
© 2022 Optica Publishing Group
Full Article | PDF ArticleMore Like This
Tongfei Lv, Yu Zhang, Lin Luo, and Xiaorong Gao
Appl. Opt. 61(9) 2219-2229 (2022)
Xianwei Ge, Shukai Chi, Wei Jia, and Ke Jiang
Appl. Opt. 61(24) 7108-7118 (2022)
Shiyuan Liu, Jingfan Fan, Dengpan Song, Tianyu Fu, Yucong Lin, Deqiang Xiao, Hong Song, Yongtian Wang, and Jian Yang
Biomed. Opt. Express 13(5) 2707-2727 (2022)