3D Object Detection for Self-Driving Cars Using Video and LiDAR: An Ablation Study Articles uri icon

authors

  • Salmane, Pascal Housam
  • Rivera Velazquez, Josue Manuel
  • Khoudour, Louahdi
  • Mai, Nguyen Anh Minh
  • Duthon, Pierre
  • Crouzil, Alain
  • Pierre, Guillaume Saint
  • VELASTIN CARROZA, SERGIO ALEJANDRO

publication date

  • March 2023

issue

  • 6

volume

  • 23

International Standard Serial Number (ISSN)

  • 1424-3210

Electronic International Standard Serial Number (EISSN)

  • 1424-8220

abstract

  • Methods based on 64-beam LiDAR can provide very precise 3D object detection. However, highly accurate LiDAR sensors are extremely costly: a 64-beam model can cost approximately USD 75,000. We previously proposed SLS¿Fusion (sparse LiDAR and stereo fusion) to fuse low-cost four-beam LiDAR with stereo cameras that outperform most advanced stereo¿LiDAR fusion methods. In this paper, and according to the number of LiDAR beams used, we analyzed how the stereo and LiDAR sensors contributed to the performance of the SLS¿Fusion model for 3D object detection. Data coming from the stereo camera play a significant role in the fusion model. However, it is necessary to quantify this contribution and identify the variations in such a contribution with respect to the number of LiDAR beams used inside the model. Thus, to evaluate the roles of the parts of the SLS¿Fusion network that represent LiDAR and stereo camera architectures, we propose dividing the model into two independent decoder networks. The results of this study show that¿starting from four beams¿increasing the number of LiDAR beams has no significant impact on the SLS¿Fusion performance. The presented results can guide the design decisions by practitioners.

keywords

  • 3d object detection; autonomous vehicle; fusion; lidar; stereo camera