Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • Comment
  • Published:

LiDAR and cameras in autonomous driving

Autonomous vehicles rely on both LiDAR and cameras for perception, with each technology offering unique advantages — cameras provide rich contextual information, whereas LiDAR delivers precise depth data. Understanding their trade-offs is crucial for creating reliable and efficient autonomous vehicles.

This is a preview of subscription content, access via your institution

Access options

Buy this article

USD 39.95

Prices may be subject to local taxes which are calculated during checkout

References

  1. Liu, F., Lu, Z. & Lin, X. Vision-based environmental perception for autonomous driving. Proc. Inst. Mech. Eng. D. 239, 39–69 (2025).

    Google Scholar 

  2. Pandharipande, A. et al. Sensing and machine learning for automotive perception: A review. IEEE Sensors J. 23, 11097–11115 (2023).

    Article  Google Scholar 

  3. Zhang, Y., Carballo, A., Yang, H. & Takeda, K. Perception and sensing for autonomous vehicles under adverse weather conditions: a survey. ISPRS J. Photogramm. Remote Sens. 196, 146–177 (2023).

    Article  Google Scholar 

  4. Li, Y. & Ibanez-Guzman, J. Lidar for autonomous driving: the principles, challenges, and trends for automotive lidar and perception systems. IEEE Signal Process. Mag. 37, 50–61 (2020).

    Google Scholar 

  5. Berrio, J. S., Shan, M., Worrall, S. & Nebot, E. Camera-LIDAR integration: probabilistic sensor fusion for semantic mapping. IEEE Trans. Intell. Transp. Syst. 23, 7637–7652 (2021).

    Article  Google Scholar 

  6. Sun, C. et al. Toward ensuring safety for autonomous driving perception: standardization progress, research advances, and perspectives. IEEE Trans. Intell. Transp. Syst. 25, 3286–3304 (2023).

    Article  Google Scholar 

  7. Li, Y., Moreau, J. & Ibanez-Guzman, J. Emergent visual sensors for autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 24, 4716–4737 (2023).

    Article  Google Scholar 

  8. Wen, L. H. & Jo, K. H. Fast and accurate 3D object detection for lidar-camera-based autonomous vehicles using one shared voxel-based backbone. IEEE Access 9, 22080–22089 (2021).

    Article  Google Scholar 

  9. Ni, P. et al. Robust 3D semantic segmentation based on multi-phase multi-modal fusion for intelligent vehicles. IEEE Trans. Intell. Veh. 9, 1602–1614 (2023).

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Javier Ibanez-Guzman.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ibanez-Guzman, J., Li, Y. LiDAR and cameras in autonomous driving. Nat Rev Electr Eng 2, 515–516 (2025). https://doi.org/10.1038/s44287-025-00176-4

Download citation

  • Published:

  • Version of record:

  • Issue date:

  • DOI: https://doi.org/10.1038/s44287-025-00176-4

This article is cited by

Search

Quick links

Nature Briefing

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing