Skip to main content
Log in

D3NET (divide and detect drivable area net): deep learning based drivable area detection and its embedded application

  • Original Research Paper
  • Published:
Journal of Real-Time Image Processing Aims and scope Submit manuscript

Abstract

Drivable area detection is an important component of various levels of autonomous driving starting from advanced driver assistance systems (ADAS) to fully automated vehicles. A drivable area detection system detects the road segment in front of the vehicle for it to drive freely and safely. Using LIght Detection And Ranging (LIDAR) or cameras, these systems need to identify areas free of vehicles, pedestrians and other objects constituting as obstacles for the vehicles movement. As such areas can vary from asphalt to dirt road with or without lane markings and with many obstacle configurations, learning-based approaches have provided effective algorithms using large training data. While accuracy is of high importance, training and runtime complexity of these methods also matter. In this work, we propose a deep learning-based method that detects the drivable area from a single image providing comparable performance with improved training and runtime performance. The model splits the given image in thin slices which are processed by a simple convolutional network regressor to model the drivable with a single parameter. The experiments on benchmark data shows comparable accuracy against the literature while showing improvement in runtime performance. It shows 237 fps operating speed and 92.55% detection performance on a Titan XP GPU while providing similar detection performance at above 30 fps on a low cost Jetson Nano module. Our code is available at https://github.com/Acuno41/D3NET.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

References

  1. Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler, M., Benenson, R., Franke, U., Roth, S., Schiele, B.: The Cityscapes Dataset for semantic urban scene understanding. In: Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)

  2. Thakur, R.: Scanning LIDAR in advanced driver assistance systems and beyond. IEEE Consum. Electron. Mag. 5(3), 48–54 (2016)

    Article  Google Scholar 

  3. Lee, Y., Park, S.: A Deep Learning-Based Perception Algorithm using 3D LiDAR for autonomous driving: simultaneous segmentation and detection network (SSADNet). Appl. Sci. 10(13), 4486 (2020)

    Article  Google Scholar 

  4. Lyu, Y., Bai, L., Huang, X.: ChipNet: real-time LiDAR processing for drivable region segmentation on an FPGA. IEEE Trans. Circuits Syst. I Regul. Pap. 66, 1769–1779 (2019)

    Article  Google Scholar 

  5. Gao, B., Xu, A., Pan, Y., Zhao, X., Yao, W., Zhao, H.: Off-road drivable area extraction using 3D LiDAR data. In: IEEE Intelligent Vehicles Symposium (IV), pp. 1505–1511 (2019)

  6. Lyu, Y., Bai, L., Huang, X.: Real-time road segmentation using LiDAR data processing on an FPGA. In: 2018 IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, pp. 1–5 (2018)

  7. Fritsch, J., Kuehnl, T., Geiger, A.: A new performance measure and evaluation benchmark for road detection algorithms. In: IEEE International Conference on Intelligent Transportation Systems (ITSC) (2013)

  8. Liu, Z., Yu, S., Zheng, N.: A co-point mapping-based approach to drivable area detection for self-driving cars. Engineering 4(4), 479–490 (2018)

    Article  Google Scholar 

  9. Liu, Z., Yu, S., Wang, X., Zheng, N.: Detecting drivable area for self-driving cars: an unsupervised approach. arXiv preprint arXiv:1750.0451 (2017)

  10. Ragurman, S.J., Park, J.: Intelligent drivable area detection system using camera and LIDAR sensor for autonomous vehicle. In: 2020 IEEE International Conference on Electro Information Technology (EIT), Chicago, IL, USA, pp. 429-436 (2020)

  11. Li, Q., Chen, L., Li, M., Shaw, S., Nüchter, A.: A sensor-fusion drivable-region and lane-detection system for autonomous vehicle navigation in challenging road scenarios. IEEE Trans. Veh. Technol. 63(2), 540–555 (2014)

    Article  Google Scholar 

  12. Poudel, R.P.K., Liwicki, S., Cipolla, R.: Fast-SCNN: fast semantic segmentation network. arXiv preprint arXiv:1902.04502 (2019)

  13. Emara, T., Munim, H.E.A.E., Abbas, H.M.: LiteSeg: a novel lightweight ConvNet for semantic segmentation. In: 2019 Digital Image Computing: Techniques and Applications (DICTA), pp. 1–7 (2019)

  14. Chen, X., Lou, X., Bai, L., Han, J.: Residual pyramid learning for single-shot semantic segmentation. IEEE Trans. Intell. Transp. Syst. 21(7), 2990–3000 (2020)

    Article  Google Scholar 

  15. Lo, S.Y., Hang, H.M., Chan, S.W., Lin, J.J.: Efficient dense modules of asymmetric convolution for real-time semantic segmentation. In: Proceedings of the ACM Multimedia Asia (MMAsia ’19), vol. 1, pp. 1–6 (2019)

  16. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.: MobileNetV2: inverted residuals and linear bottlenecks. in: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4510–4520 (2018)

  17. Brostow, G.J., Fauqueur, J., Cipolla, R.: Semantic object classes in video: a high-definition ground truth database. Pattern Recogn. Lett. 30(2), 88–97 (2009)

    Article  Google Scholar 

  18. Chen, W., Gong, X., Liu, X., Zhang, Q., Li, Y., Wang, Z.: FasterSeg: searching for faster real-time semantic segmentation. In: International Conference on Learning Representations (ICLR) (2020)

  19. Li, X., Zhou, Y., Pan, Z., Feng, J.: Partial order pruning: for best speed/accuracy trade-off in neural architecture. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9137–9145 (2019)

  20. Mazzini, D., Schettini, R.: Spatial sampling network for fast scene understanding. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1286–1296 (2019)

  21. Mehta, S., Rastegari, M., Caspi, A., Shapiro, L, Hajishirzi, H.: ESPNet: efficient spatial pyramid of dilated convolutions for semantic segmentation. arXiv preprint arXiv:1803.06815 (2018)

  22. Lo, S.Y., Hang, H.M., Chan, S.W., Lin, J.J.: Efficient dense modules of asymmetric convolution for real-time semantic segmentation. In: Proceedings of the ACM Multimedia Asia (MMAsia ’19), vol. 1, pp. 1–6 (2019)

  23. Wu, T., Tang, S., Zhang, R., Zhang, Y.: Cgnet: light-weight context guided network for semantic segmentation. arXiv preprint arXiv:1811.08201 (2019)

  24. Poudel, R.P., Bonde, U., Liwicki, S., Zach, C.: Contextnet: exploring context and detail for semantic segmentation in real-time. In: British Machine Vision Conference (BMVC) (2018)

  25. Orsic, M., Kreso, I., Bevandic, P., Segvic, S.: In defense of pre-trained Imagenet architectures for real-time semantic segmentation of road-driving images. In: 2019 IEEE/CVF conference on computer vision and pattern recognition, pp. 12607–12607 (2019)

  26. Yang, M.Y., Kumaar, S., Lyu, Y., Nex, F.: Real-time semantic segmentation with context aggregation network. arXiv preprint arXiv:2011.00993 (2021)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ayhan Küçükmanisa.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Acun, O., Küçükmanisa, A., Genç, Y. et al. D3NET (divide and detect drivable area net): deep learning based drivable area detection and its embedded application. J Real-Time Image Proc 20, 16 (2023). https://doi.org/10.1007/s11554-023-01279-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11554-023-01279-7

Keywords

Navigation