Log in

Role of Simulated Lidar Data for Training 3D Deep Learning Models: An Exhaustive Analysis

  • Research Article
  • Published:
Journal of the Indian Society of Remote Sensing Aims and scope Submit manuscript

Abstract

The use of 3D Deep Learning (DL) models for LiDAR data segmentation has attracted much interest in recent years. However, the generation of labeled point cloud data, which is a prerequisite for training DL models, is a highly resource-intensive exercise. Simulated LiDAR data, which are already labeled, provide a cost-effective alternative, but their efficacy and usefulness must be evaluated. This paper examines the role of simulated LiDAR point clouds in training DL models. A high-fidelity 3D terrain model representing the real environment is developed, and the in-house physics-based simulator “Limulator” is used to generate labeled point clouds through various realizations. The paper outlines a few major hypotheses to assess the usefulness of simulated data in training DL models. The hypotheses are designed to assess the role of simulated data alone or in combination with real data or by strategic boosting of minor classes in simulated data. Several experiments are carried out to test these hypotheses. An experiment involves training a DL model, PointCNN in this case, using various combinations of simulated and real LiDAR data and measuring its performance to segment the test data. Results show that training using simulated data alone can produce an overall accuracy (OA) of 89% and the weighted-averaged F1 score of 88.81%. It is further observed that training using a combination of simulated and real data can achieve accuracies comparable to when only a large quantity of real data is employed. Strategic boosting of minor classes in simulated data improves the accuracies of minor classes by up to 23% compared to only real data. Training a DL model using simulated data, due to the ease in its generation and positive impact on segmentation accuracy, can be highly beneficial in the use of DL for LiDAR data. The use of simulated data for training has the potential to minimize the resource-intensive exercise of develo** labeled real data.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price includes VAT (Germany)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12

Data Availability

The datasets generated during the study (simulated data) are not publicly available. However, the data can be available from the authors upon reasonable request. Furthermore, the secondary data (Real data) supporting the findings of this study are included in this published article.

Notes

  1. https://www.blender.org/

References

  • Abu Alhaija, H., Mustikovela, S. K., Mescheder, L., Geiger, A., & Rother, C. (2018). Augmented reality meets computer vision: Efficient data generation for urban driving scenes. International Journal of Computer Vision, 126(9), 961–972. https://doi.org/10.1007/s11263-018-1070-x

    Article  Google Scholar 

  • Adams, H. (2023). Adams. Hexagon. https://hexagon.com/products/product-groups/computer-aided-engineering-software/adams. Accessed 4 June 2023

  • Bousmalis, K., Silberman, N., Dohan, D., Erhan, D., & Krishnan, D. (2017). Unsupervised pixel-level domain adaptation with generative adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3722–3731)

  • Chen, C., Seff, A., Kornhauser, A., & **ao, J. (2015). Deepdriving: Learning affordance for direct perception in autonomous driving. In Proceedings of the IEEE international conference on computer vision (pp. 2722–2730)

  • Dayal, S., Goel, S., Lohani, B., Mittal, N., & Mishra, R. K. (2021). Comprehensive airborne laser scanning (ALS) simulation. Journal of the Indian Society of Remote Sensing, 49(7), 1603–1622.

    Article  Google Scholar 

  • de Melo, C. M., Rothrock, B., Gurram, P., Ulutan, O., & Manjunath, B. S. (2020). Vision-based gesture recognition in human-robot teams using synthetic data. In 2020 IEEE/RSJ International conference on intelligent robots and systems (IROS) (pp. 10278–10284). https://doi.org/10.1109/IROS45743.2020.9340728

  • de Melo, C. M., Torralba, A., Guibas, L., DiCarlo, J., Chellappa, R., & Hodgins, J. (2022). Next-generation deep learning based on simulators and synthetic data. Trends in Cognitive Sciences, 26(2), 174–187. https://doi.org/10.1016/j.tics.2021.11.008

    Article  Google Scholar 

  • Dosovitskiy, A., Ros, G., Codevilla, F., Lopez, A., & Koltun, V. (2017). CARLA: An open urban driving simulator. In Conference on robot learning (pp. 1–16). PMLR

  • Fang, J., Yan, F., Zhao, T., Zhang, F., Zhou, D., Yang, R., et al. (2018). Simulating LIDAR point cloud for autonomous driving using real-world scenes and traffic flows. https://openreview.net/forum?id=rJMw747l_4

  • Fang, J., Zhou, D., Yan, F., Zhao, T., Zhang, F., Ma, Y., et al. (2020). Augmented lidar simulator for autonomous driving. IEEE Robotics and Automation Letters, 5(2), 1931–1938.

    Article  Google Scholar 

  • Fawaz, H. I., Forestier, G., Weber, J., Idoumghar, L., & Muller, P.-A. (2018). Data augmentation using synthetic data for time series classification with deep residual networks. ar**v preprint retrieved from ar**v:1808.02455

  • Foretellix. (2023). Foretellix—Automated driving systems testing tools for simulations. Foretellix. https://www.foretellix.com/. Accessed 4 June 2023

  • Gaidon, A., Wang, Q., Cabon, Y., & Vig, E. (2016). Virtual worlds as proxy for multi-object tracking analysis. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4340–4349)

  • Gusmão, G. F., Barbosa, C. R. H., & Raposo, A. B. (2020). Development and validation of LiDAR sensor simulators based on parallel raycasting. Sensors, 20(24), 7186.

    Article  Google Scholar 

  • He, R., Sun, S., Yu, X., Xue, C., Zhang, W., Torr, P., et al. (2022). Is synthetic data from generative models ready for image recognition? ar**v preprint retrieved from ar**v:2210.07574

  • IPG, A. (2023). CarMaker|IPG Automotive. https://ipg-automotive.com/en/products-solutions/software/carmaker. Accessed 1 June 2023

  • Jaipuria, N., Zhang, X., Bhasin, R., Arafa, M., Chakravarty, P., Shrivastava, S., et al. (2020). Deflating dataset bias using synthetic data augmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (pp. 772–773)

  • Jordon, J., Yoon, J., & Van Der Schaar, M. (2018). PATE-GAN: Generating synthetic data with differential privacy guarantees. In International conference on learning representations

  • Khosla, C., & Saini, B. S. (2020). Enhancing performance of deep learning models with different data augmentation techniques: A survey. In 2020 International Conference on Intelligent Engineering and Management (ICIEM) (pp. 79–85). IEEE

  • Kumar, B., Pandey, G., Lohani, B., & Misra, S. C. (2019). A multi-faceted CNN architecture for automatic classification of mobile LiDAR data and an algorithm to reproduce point cloud samples for enhanced training. ISPRS Journal of Photogrammetry and Remote Sensing, 147, 80–89.

    Article  Google Scholar 

  • Li, Y., Bu, R., Sun, M., Wu, W., Di, X., & Chen, B. (2018). Pointcnn: Convolution on x-transformed points. Advances in neural information processing systems, 31

  • Lohani, B., & Mishra, R. K. (2007). Generating LiDAR data in laboratory: LiDAR simulator. Int. Arch. Photogramm. Remote Sens, 52(01), 264–269.

    Google Scholar 

  • Manivasagam, S., Wang, S., Wong, K., Zeng, W., Sazanovich, M., Tan, S., et al. (2020). Lidarsim: Realistic lidar simulation by leveraging the real world. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 11167–11176).

  • Morse. (2023). Modular open robots simulation engine by morse-simulator. http://morse-simulator.github.io/. Accessed 4 June 2023

  • Movshovitz-Attias, Y., Kanade, T., & Sheikh, Y. (2016). How useful is photo-realistic rendering for visual learning? In Computer Vision–ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8–10 and 15–16, 2016, Proceedings, Part III 14 (pp. 202–217). Springer.

  • Negrut, D., Serban, R., & Elmquist, A. (2020). Physics-based sensor models for virtual simulation of connected and autonomous vehicles. https://openreview.net/forum?id=rJMw747l_4

  • Nikolenko, S. I. (2021). Synthetic data for deep learning (Vol. 174). Springer International Publishing. https://doi.org/10.1007/978-3-030-75178-4

    Book  Google Scholar 

  • NVIDIA. (2023). NVIDIA DRIVE Sim Powered by Omniverse. NVIDIA. https://www.nvidia.com/en-us/self-driving-cars/simulation/. Accessed 4 June 2023

  • Park, T., Efros, A. A., Zhang, R., & Zhu, J.-Y. (2020). Contrastive learning for unpaired image-to-image translation. In A. Vedaldi, H. Bischof, T. Brox, & J.-M. Frahm (Eds.), Computer vision—ECCV 2020 (pp. 319–345). Cham: Springer International Publishing. https://doi.org/10.1007/978-3-030-58545-7_19

    Chapter  Google Scholar 

  • Prakash, A., Boochoon, S., Brophy, M., Acuna, D., Cameracci, E., State, G., et al. (2019). Structured domain randomization: bridging the reality gap by context-aware synthetic data. In 2019 International Conference on Robotics and Automation (ICRA) (pp. 7249–7255). https://doi.org/10.1109/ICRA.2019.8794443

  • Ravuri, S., & Vinyals, O. (2019). Seeing is not necessarily believing: Limitations of biggans for data augmentation. International Conference on Learning Representations. https://openreview.net/forum?id=rJMw747l_4

  • Reitmann, S., Neumann, L., & Jung, B. (2021). BLAINDER—a blender AI add-on for generation of semantically labeled depth-sensing data. Sensors, 21(6), 2144.

    Article  Google Scholar 

  • Richter, S. R., AlHaija, H. A., & Koltun, V. (2022). Enhancing photorealism enhancement. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(2), 1700–1715.

    Article  Google Scholar 

  • Rong, G., Shin, B. H., Tabatabaee, H., Lu, Q., Lemke, S., Možeiko, M., et al. (2020). Lgsvl simulator: A high fidelity simulator for autonomous driving. In 2020 IEEE 23rd International conference on intelligent transportation systems (ITSC) (pp. 1–6). IEEE

  • Ros, G., Sellart, L., Materzynska, J., Vazquez, D., & Lopez, A. M. (2016). The SYNTHIA Dataset: A large collection of synthetic images for semantic segmentation of urban scenes. In 2016 IEEE conference on computer vision and pattern recognition (CVPR) (pp. 3234–3243). Presented at the 2016 IEEE conference on computer vision and pattern recognition (CVPR). https://doi.org/10.1109/CVPR.2016.352

  • Sallab, A. E., Sobh, I., Zahran, M., & Essam, N. (2019a). LiDAR Sensor modeling and data augmentation with GANs for autonomous driving. ar**v preprint retrieved from ar**v:1905.07290

  • Sallab, A. E., Sobh, I., Zahran, M., & Shawky, M. (2019b). Unsupervised neural sensor models for synthetic lidar data augmentation. ar**v preprint retrieved from ar**v:1911.10575

  • Shafaei, A., Little, J., & Schmidt, M. (2016). Play and learn: Using video games to train computer vision models. In Procedings of the British machine vision conference 2016 (pp. 26.1–26.13). Presented at the British machine vision conference 2016, York, UK: British Machine Vision Association. https://doi.org/10.5244/C.30.26

  • Shah, S., Dey, D., Lovett, C., & Kapoor, A. (2018). Airsim: High-fidelity visual and physical simulation for autonomous vehicles. Field and service robotics (pp. 621–635). Cham: Springer.

    Chapter  Google Scholar 

  • Shrivastava, A., Pfister, T., Tuzel, O., Susskind, J., Wang, W., & Webb, R. (2017). Learning from simulated and unsupervised images through adversarial training. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2107–2116).

  • Singer, N. M., & Asari, V. K. (2021). Dales objects: A large scale benchmark dataset for instance segmentation in aerial lidar. IEEE Access, 9, 97495–97504.

    Article  Google Scholar 

  • Tiago, C., Gilbert, A., Beela, A. S., Aase, S. A., Snare, S. R., Šprem, J., & McLeod, K. (2022). A data augmentation pipeline to generate synthetic labeled datasets of 3D echocardiography images using a GAN. IEEE Access, 10, 98803–98815. https://doi.org/10.1109/ACCESS.2022.3207177

    Article  Google Scholar 

  • Trabucco, B., Doherty, K., Gurinas, M., & Salakhutdinov, R. (2023). Effective data augmentation with diffusion models. ar**v preprint retrieved from ar**v:2302.07944

  • Tran, T., Pham, T., Carneiro, G., Palmer, L., & Reid, I. (2017). A bayesian data augmentation approach for learning deep models. Advances in neural information processing systems, 30

  • Tremblay, J., Prakash, A., Acuna, D., Brophy, M., Jampani, V., Anil, C., et al. (2018). Training deep networks with synthetic data: bridging the reality gap by domain randomization. In 2018 IEEE/CVF conference on computer vision and pattern recognition workshops (CVPRW) (pp. 1082–10828). Presented at the 2018 IEEE/CVF conference on computer vision and pattern recognition workshops (CVPRW). https://doi.org/10.1109/CVPRW.2018.00143

  • van Breugel, B., Kyono, T., Berrevoets, J., & van der Schaar, M. (2021). Decaf: Generating fair synthetic data using causally-aware generative networks. Advances in Neural Information Processing Systems, 34, 22221–22233.

    Google Scholar 

  • Wang, F., Zhuang, Y., Gu, H., & Hu, H. (2019). Automatic generation of synthetic LiDAR point clouds for 3-D data analysis. IEEE Transactions on Instrumentation and Measurement, 68(7), 2671–2673.

    Article  Google Scholar 

  • Winiwarter, L., Pena, A. M. E., Weiser, H., Anders, K., Sanchez, J. M., Searle, M., & Höfle, B. (2021). Virtual laser scanning with HELIOS++: A novel take on ray tracing-based simulation of topographic 3D laser scanning. ar**v preprint retrieved from ar**v:2101.09154.

  • **ao, A., Huang, J., Guan, D., Zhan, F., & Lu, S. (2022). Transfer learning from synthetic to real LiDAR point cloud for semantic segmentation. In Proceedings of the AAAI conference on artificial intelligence (vol. 36, pp. 2795–2803).

  • Xu, Y., Noy, A., Lin, M., Qian, Q., Li, H., & **, R. (2020). Wemix: How to better utilize data augmentation. ar**v preprint retrieved from ar**v:2010.01267

  • Ye, Z., Xu, Y., Huang, R., Tong, X., Li, X., Liu, X., et al. (2020). Lasdu: A large-scale aerial lidar dataset for semantic labeling in dense urban areas. ISPRS International Journal of Geo-Information, 9(7), 450.

    Article  Google Scholar 

  • Yue, X., Wu, B., Seshia, S. A., Keutzer, K., & Sangiovanni-Vincentelli, A. L. (2018). A lidar point cloud generator: from a virtual world to autonomous driving. In Proceedings of the 2018 ACM on international conference on multimedia retrieval (pp. 458–464).

  • Zhang, X., Chen, R., Li, A., **ang, F., Qin, Y., Gu, J., et al. (2023). Close the optical sensing domain gap by physics-grounded active stereo sensor simulation. IEEE transactions on robotics

  • Zhu, J. -Y., Park, T., Isola, P., & Efros, A. A. (2017). Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 2223–2232).

Download references

Acknowledgements

The authors used the support provided by the cluster project under the Data Science (DS) Research of Frontier and Futuristic Technologies (FFT) and National Geospatial Programme (NGP) Division of the Department of Science and Technology (DST), Government of India, New Delhi.

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the study conception and design. Bharat Lohani has supervised the experiments and the work to compile the manuscript. 3D models were created by Parvej Khan and simulated LiDAR data were generated by Siddhartha Gupta. The first draft of the manuscript was written by Parvej Khan and Vaibhav Kumar and all authors commented on previous versions of the manuscript. Experiments were performed by Parvej Khan and all the authors performed thorough discussion and analysis of the outcomes.

Corresponding author

Correspondence to Parvej Khan.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Ethical Approval and Informed Consent

This article does not contain any studies involving animals performed by any of the authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lohani, B., Khan, P., Kumar, V. et al. Role of Simulated Lidar Data for Training 3D Deep Learning Models: An Exhaustive Analysis. J Indian Soc Remote Sens (2024). https://doi.org/10.1007/s12524-024-01905-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s12524-024-01905-2

Keywords

Navigation