Enhancing Navigational Performance with Holistic Deep-Reinforcement-Learning

  • Conference paper
  • First Online:
Intelligent Autonomous Systems 18 (IAS 2023)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 795))

Included in the following conference series:

  • 71 Accesses

Abstract

Deep Reinforcement Learning (DRL) has gained increasing popularity in the robotics community for tasks such as gras**, manipulation, and decision making. In recent years, DRL has been utilized for autonomous navigation as well—with remarkable results. However, most research works either focus on providing an end-to-end DRL approach or conduct an isolated training and integrate the DRL agent as a separate entity later on. This in turn, comes along with a number of problems such as catastrophic forgetfulness, inefficient navigation behavior, and non-optimal synchronization between different entities of the navigation stack. To address these issues, we propose a holistic Deep Reinforcement Learning training approach in which the training procedure is involving all entities of the navigation stack. This should enhance the synchronization between- and understanding of all entities of the navigation stack and as a result, improve navigational performance in crowded environments. We developed six agents with different observation spaces to study the impact of different input on the navigation behavior of the agent. We did extensive evaluations against multiple learning-based and classic model-based navigation approaches, our proposed agent could outperform the baselines in terms of efficiency and safety attaining shorter path lengths, less roundabout paths, and less collisions especially in situations with a high number of pedestrians.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now
Chapter
EUR 29.95
Price includes VAT (Germany)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
EUR 160.49
Price includes VAT (Germany)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
EUR 213.99
Price includes VAT (Germany)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Marder-Eppstein, E., Berger, E., Foote, T., Gerkey, B., Konolige, K.: The office marathon: robust navigation in an indoor office environment. In: IEEE International Conference on Robotics and Automation, pp. 300–307. IEEE (2010)

    Google Scholar 

  2. Kästner, L., Cox, J., Buiyan, T., Lambrecht, J.: All-in-one: a drl-based control switch combining state-of-the-art navigation planners. In: International Conference on Robotics and Automation (ICRA), pp. 2861–2867 (2022)

    Google Scholar 

  3. Qian, K., Ma, X., Dai, X., Fang, F.: Socially acceptable pre-collision safety strategies for human-compliant navigation of service robots. Adv. Robot. 24(13), 1813–1840 (2010)

    Google Scholar 

  4. Kästner, L., Buiyan, T., Zhao, X., Jiao, L., Shen, Z., Lambrecht, J.: Towards deployment of deep-reinforcement-learning-based obstacle avoidance into conventional autonomous navigation systems. ar**v preprint ar**v:2104.03616 (2021)

  5. Dugas, D., Nieto, J., Siegwart, R., Chung, J.J.: Navrep: unsupervised representations for reinforcement learning of robot navigation in dynamic human environments. ar**v preprint ar**v:2012.04406 (2020)

  6. Chen, C., Liu, Y., Kreiss, S., Alahi, A.: Crowd-robot interaction: crowd-aware robot navigation with attention-based deep reinforcement learning. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 6015–6022. IEEE (2019)

    Google Scholar 

  7. Faust, A., Oslund, K., Ramirez, O., Francis, A., Tapia, L., Fiser, M., Davidson, J.: Prm-rl: long-range robotic navigation tasks by combining reinforcement learning and sampling-based planning. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 5113–5120. IEEE (2018)

    Google Scholar 

  8. **ao, X., Liu, B., Warnell, G., Stone, P.: Motion planning and control for mobile robot navigation using machine learning: a survey. In: Autonomous Robots, pp. 1–29 (2022)

    Google Scholar 

  9. Brito, B., Everett, M., How, J.P., Alonso-Mora, J.: Where to go next: learning a subgoal recommendation policy for navigation in dynamic environments. IEEE Robot. Autom. Lett. 6(3), 4616–4623 (2021)

    Google Scholar 

  10. Kästner, L., Zhao, X., Buiyan, T., Li, J., Shen, Z., Lambrecht, J., Marx, C.: Connecting deep-reinforcement-learning-based obstacle avoidance with conventional global planners using waypoint generators. In: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1213–1220. IEEE

    Google Scholar 

  11. Chiang, H.-T.L., Faust, A., Fiser, M., Francis, A.: Learning navigation behaviors end-to-end with autorl. IEEE Robot. Autom. Lett. 4(2), 2007–2014 (2019)

    Google Scholar 

  12. Yavas, U., Kumbasar, T., Ure, N.K.: Model-based reinforcement learning for advanced adaptive cruise control: a hybrid car following policy. In: IEEE Intelligent Vehicles Symposium (IV), pp. 1466–1472. IEEE (2022)

    Google Scholar 

  13. Atoui, H., Sename, O., Milanés, V., Martinez, J.J.: Intelligent control switching for autonomous vehicles based on reinforcement learning. In: IEEE Intelligent Vehicles Symposium (IV), pp. 792–797. IEEE (2022)

    Google Scholar 

  14. Everett, M., Chen, Y.F., How, J.P.: Motion planning among dynamic, decision-making agents with deep reinforcement learning. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3052–3059. IEEE (2018)

    Google Scholar 

  15. LaValle, S.M., et al.: Rapidly-exploring random trees: a new tool for path planning (1998)

    Google Scholar 

  16. Hart, P.E., Nilsson, N.J., Raphael, B.: A formal basis for the heuristic determination of minimum cost paths. IEEE Trans. Syst. Sci. Cybern. 4(2), 100–107 (1968)

    Google Scholar 

  17. Güldenring, R., Görner, M., Hendrich, N., Jacobsen, N.J., Zhang, J.: Learning local planners for human-aware navigation in indoor environments

    Google Scholar 

  18. Regier, P., Gesing, L., Bennewitz, M.: Deep reinforcement learning for navigation in cluttered environments (2020)

    Google Scholar 

  19. Bansal, S., Tolani, V., Gupta, S., Malik, J., Tomlin, C.: Combining optimal control and learning for visual navigation in novel environments. In: Conference on Robot Learning. PMLR, pp. 420–429 (2020)

    Google Scholar 

  20. Kästner, L., Zhao, X., Shen, Z., Lambrecht, J.: Obstacle-aware waypoint generation for long-range guidance of deep-reinforcement-learning-based navigation approaches. ar**v preprint ar**v:2109.11639 (2021)

  21. Rösmann, C., Hoffmann, F., Bertram, T.: Planning of multiple robot trajectories in distinctive topologies. In: 2015 European Conference on Mobile Robots (ECMR), pp. 1–6. IEEE (2015)

    Google Scholar 

  22. Khatib, O.: Real-time obstacle avoidance for manipulators and mobile robots. In: Autonomous Robot Vehicles, pp. 396–404. Springer, Berlin (1986)

    Google Scholar 

  23. Rösmann, C., Makarow, A., Bertram, T.: Online motion planning based on nonlinear model predictive control with non-Euclidean rotation groups. ar**v preprint ar**v:2006.03534 (2020)

  24. Rösmann, C., Hoffmann, F., Bertram, T.: Timed-elastic-bands for time-optimal point-to-point nonlinear model predictive control. In: European Control Conference (ECC), pp. 3352–3357. IEEE (2015)

    Google Scholar 

  25. Rösmann, C.: Time-optimal nonlinear model predictive control. Ph.D. dissertation, Technische Universität Dortmund (2019)

    Google Scholar 

  26. Helbing, D., Molnar, P.: Social force model for pedestrian dynamics. Phys. Rev. E 51(5), 4282 (1995)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Linh Kästner .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Meusel, M., Kästner, L., Bhuiyan, T., Lambrecht, J. (2024). Enhancing Navigational Performance with Holistic Deep-Reinforcement-Learning. In: Lee, SG., An, J., Chong, N.Y., Strand, M., Kim, J.H. (eds) Intelligent Autonomous Systems 18. IAS 2023. Lecture Notes in Networks and Systems, vol 795. Springer, Cham. https://doi.org/10.1007/978-3-031-44851-5_5

Download citation

Publish with us

Policies and ethics

Navigation