Skip to main content

SAC-PER: A Navigation Method Based on Deep Reinforcement Learning Under Uncertain Environments

  • Conference paper
  • First Online:
Web and Big Data (APWeb-WAIM 2022)

Abstract

In real scenarios, robots usually face dynamically changing environments, and traditional navigation methods require a predefined high-precision map, which limits the achievability of navigation in dynamic and uncertain environments. To solve this problem, this paper uses a Partially Observable Markov Decision Process (POMDP) to model the uncertain navigation planning problem and proposes a soft actor-critic with prioritized experience replay (SAC-PER) method based on multi-sensor perception to achieve efficient navigation. The method uses multi-source information fusion for environment perception and Deep Reinforcement Learning (DRL) for continuous control of navigation. The multi-source SAC-PER method can effectively avoid obstacles and enable robots to perform navigation tasks autonomously in uncertain environments without building high-precision maps. We evaluate the proposed method using Robot Operating System (ROS) and Gazebo simulator. The results demonstrate that the SAC-PER method has high efficiency and robustness in different environments, and shows good generalization ability.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Chen, J., Hu, K., Li, Y.: Research on UAV multi-point navigation algorithm based on mb-rrt*. Comput. Sci. 45(6A), 85–90 (2018)

    Google Scholar 

  2. Duong, T., Das, N., Yip, M.: Autonomous navigation in unknown environments using sparse kernel-based occupancy mapping. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 9666–9672 (2020)

    Google Scholar 

  3. Gavrilov, A.V., Lenskiy, A.: Mobile robot navigation using reinforcement learning based on neural network with short term memory. In: Advanced Intelligent Computing, pp. 210–217 (2012)

    Google Scholar 

  4. Mohanan, M., Salgoankar, A.: A survey of robotic motion planning in dynamic environments. Robot. Auton. Syst. 100, 171–185 (2018)

    Article  Google Scholar 

  5. Sarkar, M., Yan, X., Erol, B.A.: A novel search and survey technique for unmanned aerial systems in detecting and estimating the area for wildfires. Robot. Auton. Syst. 145, 103848 (2021)

    Article  Google Scholar 

  6. Shade, R., Newman, P.: Choosing where to go: Complete 3d exploration with stereo. In: 2011 IEEE International Conference on Robotics and Automation, pp. 2806–2811 (2011)

    Google Scholar 

  7. Tai, L., Paolo, G., Liu, M.: Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation. In: 2017 International Conference on Intelligent Robots and Systems (IROS), pp. 31–36 (2017)

    Google Scholar 

  8. Thrun, S.: Probabilistic robotics. Commun. ACM 45, 52–57 (2002)

    Article  Google Scholar 

  9. Toan, N.D., Woo, K.G.: Mapless navigation with deep reinforcement learning based on the convolutional proximal policy optimization network. In: IEEE International Conference on Big Data and Smart Computing, pp. 298–301 (2021)

    Google Scholar 

  10. Yan, F., Zhuang, Y., Xiao, J.: 3d prm based real-time path planning for uav in complex environment. In: 2012 IEEE International Conference on Robotics and Biomimetics (ROBIO), pp. 1135–1140 (2012)

    Google Scholar 

  11. Zhou, L., Koppel, D.: Lidar slam with plane adjustment for indoor environment. IEEE Robot. Automat. Lett. 6(4), 7073–7080 (2021)

    Article  Google Scholar 

  12. Zuo, B., Chen, J., Wang, L., Wang, Y.: A reinforcement learning based robotic navigation system. In: 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 3452–3457 (2014)

    Google Scholar 

Download references

Acknowledgment

This work is supported by Civil Aircraft Special Scientific Research Project under grant number MJ-2018-S-29.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lisong Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Wang, X., Wang, L., Shen, S., Hu, L. (2023). SAC-PER: A Navigation Method Based on Deep Reinforcement Learning Under Uncertain Environments. In: Li, B., Yue, L., Tao, C., Han, X., Calvanese, D., Amagasa, T. (eds) Web and Big Data. APWeb-WAIM 2022. Lecture Notes in Computer Science, vol 13422. Springer, Cham. https://doi.org/10.1007/978-3-031-25198-6_38

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-25198-6_38

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-25197-9

  • Online ISBN: 978-3-031-25198-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics