Laser & Optoelectronics Progress, Volume. 59, Issue 18, 1828004(2022)

3D Object Detection Based on Extremely Sparse Laser Point Cloud and RGB Images

Chao Qin1,2, Yafei Wang1, Yuchao Zhang2, and Chengliang Yin1、*
Author Affiliations
  • 1School of Mechanical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
  • 2Shanghai Intelligent and Connected Vehicle R&D Center Co., Ltd., Shanghai 201499, China
  • show less
    Figures & Tables(18)
    Structure of proposed 3D object detection algorithm
    Depth completion network
    Results of dense point cloud and sparse point cloud projected on the image respectively
    Point cloud image generated from depth map
    3D object detection network based on key point feature pyramid
    Dense depth map generated from depth completion network
    Result of sparse point cloud projection on the image
    Dense depth map generated from depth completion network
    BEV map generated from dense point cloud after aerial view projection
    Detection result on BEV map
    Display effect of object detection stereo bounding box on camera RGB pictures
    • Table 1. Target detection accuracy of proposed algorithm on KITTI dataset

      View table

      Table 1. Target detection accuracy of proposed algorithm on KITTI dataset

      AlgorithmCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
      EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
      Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    • Table 2. Target detection accuracy under the condition of sparse point cloud BEV as theinput of key point feature pyramid network

      View table

      Table 2. Target detection accuracy under the condition of sparse point cloud BEV as theinput of key point feature pyramid network

      InputCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
      EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
      Sparse point cloud4.503.152.880.960.940.90.820.780.70
      Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    • Table 3. Target detection accuracy under the condition of coded point cloud in previous view form

      View table

      Table 3. Target detection accuracy under the condition of coded point cloud in previous view form

      InputCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
      EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult

      Point cloud

      organized as(xyz

      60.2552.8045.6035.8631.3829.7241.8239.5636.09
      Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    • Table 4. Target detection accuracy under the condition of only taking the picture as the input of depth complement network

      View table

      Table 4. Target detection accuracy under the condition of only taking the picture as the input of depth complement network

      InputCar(IOU is 0.7)Person(IOU is 0.5)Bicycl e(IOU is 0.5)
      EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
      Image34.5221.0419.0322.2413.5612.265.633.433.10
      Proposed algorithm87.9877.1473.3345.9738.9435.8168.1255.2553.55
    • Table 5. Target detection accuracy under different point cloud down sampling rates

      View table

      Table 5. Target detection accuracy under different point cloud down sampling rates

      Down sampling rateCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
      EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
      1%44.3032.0328.3231.5622.3620.7916.2712.8511.50
      6%73.5460.1755.8038.9133.4029.9548.240.7639.80
      8%81.8569.8467.8041.2035.2131.9053.5247.9243.77
      10%87.9877.1473.3345.9738.9435.8168.1255.2553.55
      12%88.2078.2673.6546.0139.7036.1168.8055.7354.02
    • Table 6. Comparison of 3D object detection algorithms on KITTI dataset

      View table

      Table 6. Comparison of 3D object detection algorithms on KITTI dataset

      AlgorithmModalityCar(IOU is 0.7)Person(IOU is 0.5)Bicycle(IOU is 0.5)
      EasyModerateDifficultEasyModerateDifficultEasyModerateDifficult
      VoxelNet64-line LiDAR77.4765.1157.7339.4833.6931.5161.2248.3644.37
      SECOND64-line LiDAR83.1373.6666.2051.0742.5637.2970.5153.8546.90
      PointRCNN64-line LiDAR85.9475.7668.3249.4341.7838.6373.9359.6053.59
      SS3D30Camera10.787.686.512.311.781.482.801.451.35
      D4LCN31Camera16.6511.729.514.553.422.832.451.671.36
      AVOD64-line LiDAR+camera76.3966.4760.2336.1027.8625.7657.1942.0838.29
      Frustum PointNets64-line LiDAR+camera82.1969.7960.5950.5342.1538.0872.2756.1249.01
      Proposed algorithmSparse point cloud+camera87.9877.1473.3345.9738.9435.8168.1255.2553.55
    • Table 7. Running time comparison of 3D object detection algorithms on KITTI dataset

      View table

      Table 7. Running time comparison of 3D object detection algorithms on KITTI dataset

      ParameterVoxelNetSECONDPointRCNNSS3DD4LCNProposed algorithm
      Running time /s0.230.050.10.050.20.08
    Tools

    Get Citation

    Copy Citation Text

    Chao Qin, Yafei Wang, Yuchao Zhang, Chengliang Yin. 3D Object Detection Based on Extremely Sparse Laser Point Cloud and RGB Images[J]. Laser & Optoelectronics Progress, 2022, 59(18): 1828004

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Remote Sensing and Sensors

    Received: Aug. 3, 2021

    Accepted: Aug. 23, 2021

    Published Online: Aug. 29, 2022

    The Author Email: Yin Chengliang (clyin@sjtu.edu.cn)

    DOI:10.3788/LOP202259.1828004

    Topics