Lyapunov-Inspired Deep Reinforcement Learning for Robot Navigation in Obstacle Environments

The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL reward design inspired by Lyapunov stability theory for safe robot navigation in the presence of obstacles. The navigation problem is formulated...

Full description

Saved in:
Bibliographic Details
Published in2025 IEEE Symposium on Computational Intelligence on Engineering/Cyber Physical Systems (CIES) pp. 1 - 8
Main Authors Ugurlu, Halil Ibrahim, Redder, Adrian, Kayacan, Erdal
Format Conference Proceeding
LanguageEnglish
Published IEEE 17.03.2025
Subjects
Online AccessGet full text
DOI10.1109/CIES64955.2025.11007627

Cover

Loading…
Abstract The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL reward design inspired by Lyapunov stability theory for safe robot navigation in the presence of obstacles. The navigation problem is formulated as a state-space control problem with close obstacle locations integrated into the state representation. To ensure safe obstacle avoidance, we introduce a novel reward-shaping strategy utilizing a Lyapunov function that discourages fast movement toward obstacles. Our numerical experiments demonstrate the effectiveness of the reward design strategy compared to baselines in achieving consistent superior learning with higher mission completion rates while maintaining speeds closer to a desired target speed. In addition, we show that our reward design enables a generally smaller choice for the discount factor for value-function-based DRL algorithms, which can lead to faster convergence. This is possible since the reward design merely penalizes the one-step decay of the Lyapunov function. Furthermore, policy training simulations employ an early episode termination method to constrain exploration and add more valuable samples to the DRL training replay memory. Finally, real-world experiments with a quadrotor validate the ability of our method to safely navigate around varying densities of obstacles. The proposed method consistently takes cautious maneuvers near obstacles by slowing down, achieving greater obstacle clearance compared to baseline, although with an increase in mission completion time.
AbstractList The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL reward design inspired by Lyapunov stability theory for safe robot navigation in the presence of obstacles. The navigation problem is formulated as a state-space control problem with close obstacle locations integrated into the state representation. To ensure safe obstacle avoidance, we introduce a novel reward-shaping strategy utilizing a Lyapunov function that discourages fast movement toward obstacles. Our numerical experiments demonstrate the effectiveness of the reward design strategy compared to baselines in achieving consistent superior learning with higher mission completion rates while maintaining speeds closer to a desired target speed. In addition, we show that our reward design enables a generally smaller choice for the discount factor for value-function-based DRL algorithms, which can lead to faster convergence. This is possible since the reward design merely penalizes the one-step decay of the Lyapunov function. Furthermore, policy training simulations employ an early episode termination method to constrain exploration and add more valuable samples to the DRL training replay memory. Finally, real-world experiments with a quadrotor validate the ability of our method to safely navigate around varying densities of obstacles. The proposed method consistently takes cautious maneuvers near obstacles by slowing down, achieving greater obstacle clearance compared to baseline, although with an increase in mission completion time.
Author Ugurlu, Halil Ibrahim
Redder, Adrian
Kayacan, Erdal
Author_xml – sequence: 1
  givenname: Halil Ibrahim
  surname: Ugurlu
  fullname: Ugurlu, Halil Ibrahim
  organization: Aarhus University,Department of Electrical and Computer Engineering,Aarhus,Denmark
– sequence: 2
  givenname: Adrian
  surname: Redder
  fullname: Redder, Adrian
  organization: Paderborn University,Department of Automatic Control,Paderborn,Germany
– sequence: 3
  givenname: Erdal
  surname: Kayacan
  fullname: Kayacan, Erdal
  organization: Paderborn University,Department of Automatic Control,Paderborn,Germany
BookMark eNo1j8tKw0AYRkfQhda-geC8QOpcMpcsJUYbCBaqrlyUSfJPGWj_CZMY6NtrUVcHDpwPvhtyiRGBkHvOVpyz4qGsqzedF0qtBBPq7JjRwlyQZWEKKyVXzAojrslnc3LDF8Y5q3EcQoKePgEMdAsBfUwdHAEn2oBLGHBPfxTdxjZO9NXNYe-mEJEGpJt2nFx3AFrhHFLEczXekivvDiMs_7ggH8_Ve7nOms1LXT42WeDGTplXkknBmPAtA-OdN8pIW4A10ivfe9HZjvd52-rcCyu91jnoovMsl85IZuSC3P3uBgDYDSkcXTrt_j_LbzuwUkU
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/CIES64955.2025.11007627
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE/IET Electronic Library
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE/IET Electronic Library
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 9798331508272
EndPage 8
ExternalDocumentID 11007627
Genre orig-research
GroupedDBID 6IE
6IL
CBEJK
RIE
RIL
ID FETCH-LOGICAL-i178t-f53032002fb0e7faf757389e873f5fdf2c8c1d4bb64f283f664e69cf043a73073
IEDL.DBID RIE
IngestDate Thu May 29 05:57:27 EDT 2025
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i178t-f53032002fb0e7faf757389e873f5fdf2c8c1d4bb64f283f664e69cf043a73073
PageCount 8
ParticipantIDs ieee_primary_11007627
PublicationCentury 2000
PublicationDate 2025-March-17
PublicationDateYYYYMMDD 2025-03-17
PublicationDate_xml – month: 03
  year: 2025
  text: 2025-March-17
  day: 17
PublicationDecade 2020
PublicationTitle 2025 IEEE Symposium on Computational Intelligence on Engineering/Cyber Physical Systems (CIES)
PublicationTitleAbbrev CIES
PublicationYear 2025
Publisher IEEE
Publisher_xml – name: IEEE
Score 1.9047948
Snippet The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL...
SourceID ieee
SourceType Publisher
StartPage 1
SubjectTerms Collision avoidance
Convergence
Deep reinforcement learning
Lyapunov methods
Lyapunov theory
Navigation
Obstacle avoidance
Planning
Quadrotor navigation
Quadrotors
Reward design
Robots
Safety
Training
Title Lyapunov-Inspired Deep Reinforcement Learning for Robot Navigation in Obstacle Environments
URI https://ieeexplore.ieee.org/document/11007627
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1bS8MwFA66J59UnHgnD7626yW3Ps-NKTplOBj4MJr0RIbQDmkH-us9yTpvIPhWQkNLktPzNfm-7xBymWsJDAPaMdWSwGH0QNtCB8ALbfDrx5Uv33Y3FqMpu5nxWStW91oYAPDkMwjdpT_LLyrTuK2ynrM3w-CV22Qb19larNVytuIo6_URQAkE_Bx_-xIebu7-UTfFp43hLhlvHrhmi7yETa1D8_7Li_Hfb7RHul8KPfrwmXv2yRaUB-Tp9i1fNmW1Cq5Ld4AOBb0CWNIJeHtU43cCaeuo-kyxiU4qXdV0nK-800ZV0kVJ7zUiRlxNdPBNBNcl0-HgsT8K2uIJwSKWqg4sT11t9CixOgJpcyu5RHACSqaW28ImRpm4YFoLZhFiWCEYiMzYiKW5dIF_SDplVcIRoSxniGq5yDKIsQNXRqVKsSKRkZUiU8ek60Zmvlz7Y8w3g3LyR_sp2XET5JhcsTwjnfq1gXNM7bW-8FP6AaTopYs
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1dS8MwFA06H_RJxYnf5sHXdm2Xrz7PjU23KmODgQ-jSW9kCO2QdqC_3iTr_ALBtxJaGpLe3pPknHMRukklB2IC2jLVIs9idE_qTHpAM6nM348KV75tlLD-lNzN6KwWqzstDAA48hn49tKd5WeFquxWWcvam5ng5dtoxyR-QtdyrZq1FQZxq2MgFDOQn5qFX0T9zf0_Kqe4xNHbR8nmlWu-yItfldJX77_cGP_dpwPU_NLo4cfP7HOItiA_Qk_Dt3RZ5cXKG-T2CB0yfAuwxGNwBqnK7QXi2lP1GZsmPC5kUeIkXTmvjSLHixw_SIMZzfeEu99kcE007XUnnb5Xl0_wFiEXpadp21ZHDyItA-A61ZxyA09A8LamOtOREirMiJSMaAMyNGMEWKx0QNopt6F_jBp5kcMJwiQlBtdSFscQmgeoUKItBMkiHmjOYnGKmnZk5su1Q8Z8Myhnf7Rfo93-ZDScDwfJ_Tnas5NleV0hv0CN8rWCS5PoS3nlpvcDfxWo2A
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2025+IEEE+Symposium+on+Computational+Intelligence+on+Engineering%2FCyber+Physical+Systems+%28CIES%29&rft.atitle=Lyapunov-Inspired+Deep+Reinforcement+Learning+for+Robot+Navigation+in+Obstacle+Environments&rft.au=Ugurlu%2C+Halil+Ibrahim&rft.au=Redder%2C+Adrian&rft.au=Kayacan%2C+Erdal&rft.date=2025-03-17&rft.pub=IEEE&rft.spage=1&rft.epage=8&rft_id=info:doi/10.1109%2FCIES64955.2025.11007627&rft.externalDocID=11007627