• Laser & Optoelectronics Progress
  • Vol. 58, Issue 20, 2010006 (2021)
Haidong Zhang, Yiming Xu*, Li Wang, Chunlei Bian, and Fangjie Zhou
Author Affiliations
  • School of Electrical Engineering, Nantong University, Nantong, Jiangsu 226019, China
  • show less
    DOI: 10.3788/LOP202158.2010006 Cite this Article Set citation alerts
    Haidong Zhang, Yiming Xu, Li Wang, Chunlei Bian, Fangjie Zhou. Visual Odometry Based on Improved Dual-Stream Network Structure[J]. Laser & Optoelectronics Progress, 2021, 58(20): 2010006 Copy Citation Text show less

    Abstract

    Because conventional visual odometry (VO) has cumbersome implementation process and complex calculation problems, a VO based on an improved dual-stream network structure is proposed. The proposed VO uses a dual-stream convolutional neural network structure that can simultaneously feed RGB and depth images into the model for training, use the Inception network structure to improve the convolutional layer, and reduce the number of parameters in the convolutional layer. Simultaneously, an attention mechanism is introduced to the convolutional layer to enhance the network’s recognition of image features and the system’s robustness. After being trained and tested on the KITTI dataset, the proposed improved model is compared with the VISO2-M, VISO2-S, and SfMLearner. The results show that the proposed model’s rotation and translation errors are significantly reduced compared with VISO2-M and SfMLearner when using monocular cameras and comparable to VISO2-S when using binocular cameras.
    Haidong Zhang, Yiming Xu, Li Wang, Chunlei Bian, Fangjie Zhou. Visual Odometry Based on Improved Dual-Stream Network Structure[J]. Laser & Optoelectronics Progress, 2021, 58(20): 2010006
    Download Citation