• Laser & Optoelectronics Progress
  • Vol. 59, Issue 18, 1828004 (2022)
Chao Qin1、2, Yafei Wang1, Yuchao Zhang2, and Chengliang Yin1、*
Author Affiliations
  • 1School of Mechanical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
  • 2Shanghai Intelligent and Connected Vehicle R&D Center Co., Ltd., Shanghai 201499, China
  • show less
    DOI: 10.3788/LOP202259.1828004 Cite this Article Set citation alerts
    Chao Qin, Yafei Wang, Yuchao Zhang, Chengliang Yin. 3D Object Detection Based on Extremely Sparse Laser Point Cloud and RGB Images[J]. Laser & Optoelectronics Progress, 2022, 59(18): 1828004 Copy Citation Text show less
    Structure of proposed 3D object detection algorithm
    Fig. 1. Structure of proposed 3D object detection algorithm
    Depth completion network
    Fig. 2. Depth completion network
    Results of dense point cloud and sparse point cloud projected on the image respectively
    Fig. 3. Results of dense point cloud and sparse point cloud projected on the image respectively
    Point cloud image generated from depth map
    Fig. 4. Point cloud image generated from depth map
    3D object detection network based on key point feature pyramid
    Fig. 5. 3D object detection network based on key point feature pyramid
    Dense depth map generated from depth completion network
    Fig. 6. Dense depth map generated from depth completion network
    Result of sparse point cloud projection on the image
    Fig. 7. Result of sparse point cloud projection on the image
    Dense depth map generated from depth completion network
    Fig. 8. Dense depth map generated from depth completion network
    BEV map generated from dense point cloud after aerial view projection
    Fig. 9. BEV map generated from dense point cloud after aerial view projection
    Detection result on BEV map
    Fig. 10. Detection result on BEV map
    Display effect of object detection stereo bounding box on camera RGB pictures
    Fig. 11. Display effect of object detection stereo bounding box on camera RGB pictures
    AlgorithmCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
    EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
    Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    Table 1. Target detection accuracy of proposed algorithm on KITTI dataset
    InputCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
    EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
    Sparse point cloud4.503.152.880.960.940.90.820.780.70
    Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    Table 2. Target detection accuracy under the condition of sparse point cloud BEV as theinput of key point feature pyramid network
    InputCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
    EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult

    Point cloud

    organized as(xyz

    60.2552.8045.6035.8631.3829.7241.8239.5636.09
    Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    Table 3. Target detection accuracy under the condition of coded point cloud in previous view form
    InputCar(IOU is 0.7)Person(IOU is 0.5)Bicycl e(IOU is 0.5)
    EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
    Image34.5221.0419.0322.2413.5612.265.633.433.10
    Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    Table 4. Target detection accuracy under the condition of only taking the picture as the input of depth complement network
    Down sampling rateCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
    EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
    1%44.3032.0328.3231.5622.3620.7916.2712.8511.50
    6%73.5460.1755.8038.9133.4029.9548.240.7639.80
    8%81.8569.8467.8041.2035.2131.9053.5247.9243.77
    10%87.9877.1473.3345.9738.9435.8168.1255.2553.55
    12%88.2078.2673.6546.0139.7036.1168.8055.7354.02
    Table 5. Target detection accuracy under different point cloud down sampling rates
    AlgorithmModalityCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
    EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
    VoxelNet64-line LiDAR77.4765.1157.7339.4833.6931.5161.2248.3644.37
    SECOND64-line LiDAR83.1373.6666.2051.0742.5637.2970.5153.8546.90
    PointRCNN64-line LiDAR85.9475.7668.3249.4341.7838.6373.9359.6053.59
    SS3D30Camera10.787.686.512.311.781.482.801.451.35
    D4LCN31Camera16.6511.729.514.553.422.832.451.671.36
    AVOD64-line LiDAR+camera76.3966.4760.2336.1027.8625.7657.1942.0838.29
    Frustum PointNets64-line LiDAR+camera82.1969.7960.5950.5342.1538.0872.2756.1249.01
    Proposed algorithmSparse point cloud+camera87.9877.1473.3345.9738.9435.8168.1255.2553.55
    Table 6. Comparison of 3D object detection algorithms on KITTI dataset
    ParameterVoxelNetSECONDPointRCNNSS3DD4LCNProposed algorithm
    Running time /s0.230.050.10.050.20.08
    Table 7. Running time comparison of 3D object detection algorithms on KITTI dataset
    Chao Qin, Yafei Wang, Yuchao Zhang, Chengliang Yin. 3D Object Detection Based on Extremely Sparse Laser Point Cloud and RGB Images[J]. Laser & Optoelectronics Progress, 2022, 59(18): 1828004
    Download Citation