Professional Documents
Culture Documents
Enabling 3D Object Detection With A Low-Resolution Lidar
Enabling 3D Object Detection With A Low-Resolution Lidar
8, AUGUST 2015 1
Abstract—Light Detection And Ranging (LiDAR) has been TABLE I: A comparison of VLP-16 and HDL-64E LiDARs
widely used in autonomous vehicles for perception and local- LiDAR type VLP-16 HDL-64E
ization. However, the cost of a high-resolution LiDAR is still Channel number 16 64
prohibitively expensive, while its low-resolution counterpart is
arXiv:2105.01765v2 [cs.CV] 4 May 2022
(d) dense depth map scenario 1 (e) dense depth map scenario 2 (f) dense depth map scenario 3
(g) sparse depth map scenario 1 (h) sparse depth map scenario 2 (i) sparse depth map scenario 3
Fig. 1: Comparison of depth map from 16-line LiDAR (bottom) and 64-line LiDAR (middle) to their RGB image (top), on
which red boxes represent the short-range vehicles, orange boxes show the medium range vehicles and the far vehicles are
marked by blue boxes. Green boxes illustrate the occluded vehicles.
TABLE III: BEV performance comparison on KITTI object detection validation dataset. This table shows APBEV (in
%) of the car category, corresponding to average precision of the bird’s-eye view.
IoU=0.5 IoU=0.7
Detection Networks Input
Easy Moderate Hard Easy Moderate Hard
PIXOR[10] LiDAR only (64-line) 94.2 86.7 86.1 85.2 81.2 76.1
PIXOR LiDAR only (16-line) 60.7 51.2 46.8 53.8 47.1 39.1
Ours LiDAR (16-line) + Camera 89.0 75.8 68.1 75.4 61.2 55.2
V. C ONCLUSION
This paper presents a framework that enables 3D object
detection using a low-resolution LiDAR. By cascading a depth
completion network with an object detector, it first converts the
sparse point cloud into a denser depth map that is subsequently
processed for 3D object detection. It demonstrates 3D object
detection with only a 16-line LiDAR and a camera. When
evaluated on KITTI dataset, the proposed solution achieves
high accuracy in object detection for both easy and moderate
cases, comparable to the benchmarks using 64-line LiDARs.
R EFERENCES
[1] L. Gigli et al, ”Road Segmentation on low resolution Lidar point clouds
for autonomous vehicles,” arXiv preprint arXiv:2005.13102, 2020
[2] L. Caltagirone, S. Scheidegger, L. Svensson and M. Wahde, ”Fast LIDAR-
based road detection using fully convolutional neural networks,” 2017
IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, 2017, pp.
1019-1024
[3] A. Elhadidy, M. Afifi, M. Hassoubah, Y. Ali and M. ElHelw, ”Improved
Semantic Segmentation of Low-Resolution 3D Point Clouds Using Su-
pervised Domain Adaptation,” 2020 2nd Novel Intelligent and Leading
Emerging Sciences Conference (NILES), Giza, Egypt, 2020, pp. 588-593
[4] I. del Pino et al, ”Low Resolution Lidar-Based Multi-Object Tracking
for Driving Applications,” 2017 Iberian Robotics Conference (ROBOT),
2017, pages 287–298
[5] M. Seikavandi, K. Nasrollahi, and T. B. Moeslund, ”Deep car detection
by fusing grayscale image and weighted upsampled LiDAR depth”, Proc.
SPIE 11605, Thirteenth International Conference on Machine Vision,
1160524, 2020
[6] F. Farahnakian and J. Heikkonen, ”Fusing LiDAR and Color Imagery for
Object Detection using Convolutional Neural Networks,” 2020 IEEE 23rd
International Conference on Information Fusion (FUSION), Rustenburg,
South Africa, 2020, pp. 1-7
[7] J. Redmon and A. Farhadi, “Yolov3: An incremental improvement,” arXiv
preprint arXiv:1804.02767, 2018
[8] J. Beltrán, C. Guindel, F. M. Moreno, D. Cruzado, F. Garcı́a and A.
Fig. 5: Visualization of object detection from the proposed De La Escalera, ”BirdNet+: End-to-End 3D Object Detection in LiDAR
framework, where the green boxes are ground truth and the Bird’s Eye View,” 2020 23st International Conference on Intelligent
blue boxes represent the predicted results Transportation Systems (ITSC), Maui, HI, 2018, pp. 3517-3523
[9] A. Barrera, C. Guindel, J. Beltrán and F. Garcı́a, ”BirdNet+: End-to-
End 3D Object Detection in LiDAR Bird’s Eye View,” 2020 IEEE 23rd
International Conference on Intelligent Transportation Systems (ITSC),
Rhodes, Greece, 2020, pp. 1-6
[10] B. Yang, W. Luo and R. Urtasun, ”PIXOR: Real-time 3D Object
Detection from Point Clouds,” 2018 IEEE/CVF Conference on Computer
Vision and Pattern Recognition, Salt Lake City, UT, 2018, pp. 7652-7660
[11] F. Ma, G. V. Cavalheiro and S. Karaman, ”Self-Supervised Sparse-to-
Dense: Self-Supervised Depth Completion from LiDAR and Monocular
Camera,” 2019 International Conference on Robotics and Automation
(ICRA), Montreal, QC, Canada, 2019, pp. 3288-3295
[12] R. Qian et al., ”End-to-End Pseudo-LiDAR for Image-Based 3D Object
Detection,” 2020 IEEE/CVF Conference on Computer Vision and Pattern
Recognition (CVPR), Seattle, WA, USA, 2020, pp. 5880-5889
[13] L.C. Chen, Y. Zhu, G. Papandreou, F. Schroff,and H. Adam, ”Encoder-
decoder with atrous separable convolution for semantic image segmen-
tation,” 2018 15th European Conference on Computer Vision (ECCV),
Munich, Germany, 2018, pp. 801-818
[14] H. Liu, H. Yuan, Q. Liu, J. Hou and J. Liu, ”A Comprehensive Study
and Comparison of Core Technologies for MPEG 3-D Point Cloud
Fig. 6: Precision-Recall curve of the proposed framework on Compression,” in IEEE Transactions on Broadcasting, vol. 66, no. 3, pp.
KITTI val dataset. 701-717, September 2020