• Laser & Optoelectronics Progress
  • Vol. 57, Issue 2, 21012 (2020)
Zhao Shuanfeng, Huang Tao*, Xu Qian, and Geng Longlong
Author Affiliations
  • College of Mechanical Engineering, Xi''an University of Science and Technology, Xi''an, Shaanxi 710054, China
  • show less
    DOI: 10.3788/LOP57.021012 Cite this Article Set citation alerts
    Zhao Shuanfeng, Huang Tao, Xu Qian, Geng Longlong. Unsupervised Monocular Depth Estimation for Autonomous Flight of Drones[J]. Laser & Optoelectronics Progress, 2020, 57(2): 21012 Copy Citation Text show less
    Principle of binocular depth estimation
    Fig. 1. Principle of binocular depth estimation
    Structural diagram of unsupervised monocular depth estimation
    Fig. 2. Structural diagram of unsupervised monocular depth estimation
    Model of image reconstruction
    Fig. 3. Model of image reconstruction
    Loss function of each part of training process. (a) Structural similarity loss of reconstructed image and original image; (b) absolute value loss of difference between reconstructed image and original image; (c) total image reconstruction loss; (d) loss of disparity smoothness; (e) loss of consistency in left and right disparity maps; (f) total loss of our model
    Fig. 4. Loss function of each part of training process. (a) Structural similarity loss of reconstructed image and original image; (b) absolute value loss of difference between reconstructed image and original image; (c) total image reconstruction loss; (d) loss of disparity smoothness; (e) loss of consistency in left and right disparity maps; (f) total loss of our model
    Platform of drone experiment. (a) Drone; (b) connection of NVIDIA Jeston TX2 and Pixhawk
    Fig. 5. Platform of drone experiment. (a) Drone; (b) connection of NVIDIA Jeston TX2 and Pixhawk
    Examples of depth map predicted on KITTI dataset. (a) Input image; (b) ground truth depth map; (c) depth map predicted by Ref. [15] ; (d) depth map predicted in Ref. [20]; (e) depth map predicted by our model based on VGG-16; (f) depth map predicted by our model based on ResNet-50
    Fig. 6. Examples of depth map predicted on KITTI dataset. (a) Input image; (b) ground truth depth map; (c) depth map predicted by Ref. [15] ; (d) depth map predicted in Ref. [20]; (e) depth map predicted by our model based on VGG-16; (f) depth map predicted by our model based on ResNet-50
    Examples of depth map predicted in real outdoor scenes. (a) Input images; (b) ground truth depth maps
    Fig. 7. Examples of depth map predicted in real outdoor scenes. (a) Input images; (b) ground truth depth maps
    MethodSupervisedError (lower is better)Accuracy (higher is better)Time /s
    ERELERMSELog ERMSEδ<1.25δ<1.252δ<1.252
    Ref. [12]Yes0.2036.3070.2820.7020.8900.9580.051
    Ref. [15]Yes0.2026.5230.2750.6780.8950.9650.045
    Ref. [19]No0.2086.8560.2830.6780.8850.9570.062
    Ref. [20]No0.1595.7890.2340.7960.9230.9630.057
    Our (VGG-16)No0.1485.4960.2260.8120.9120.9600.056
    Our (RseNet-50)No0.1245.3310.2190.8470.9450.9750.048
    Table 1. Comparison of experimental results on KITTI dataset
    MethodSupervisedError (lower is better)Accuracy (higher is better)Time/s
    ERELERMSELog ERMSEδ<1.25δ<1.252δ<1.252
    Ref. [12]Yes0.4178.5260.4030.6920.8990.9480.068
    Ref. [15]Yes0.4629.9720.4560.6560.8870.9450.048
    Ref. [19]No0.4438.3260.3980.6620.8850.9320.074
    Ref. [20]No0.3877.8950.3540.7040.8990.9460.054
    Our (VGG16)No0.3618.1020.3770.7270.9050.9580.061
    Our (RseNet-50)No0.3287.5290.3480.7510.9240.9620.053
    Table 2. Comparisonof experimental results on Make3D dataset
    Zhao Shuanfeng, Huang Tao, Xu Qian, Geng Longlong. Unsupervised Monocular Depth Estimation for Autonomous Flight of Drones[J]. Laser & Optoelectronics Progress, 2020, 57(2): 21012
    Download Citation