Lyapunov-Inspired Deep Reinforcement Learning for Robot Navigation in Obstacle Environments
The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL reward design inspired by Lyapunov stability theory for safe robot navigation in the presence of obstacles. The navigation problem is formulated...
Saved in:
Published in | 2025 IEEE Symposium on Computational Intelligence on Engineering/Cyber Physical Systems (CIES) pp. 1 - 8 |
---|---|
Main Authors | , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
17.03.2025
|
Subjects | |
Online Access | Get full text |
DOI | 10.1109/CIES64955.2025.11007627 |
Cover
Loading…
Abstract | The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL reward design inspired by Lyapunov stability theory for safe robot navigation in the presence of obstacles. The navigation problem is formulated as a state-space control problem with close obstacle locations integrated into the state representation. To ensure safe obstacle avoidance, we introduce a novel reward-shaping strategy utilizing a Lyapunov function that discourages fast movement toward obstacles. Our numerical experiments demonstrate the effectiveness of the reward design strategy compared to baselines in achieving consistent superior learning with higher mission completion rates while maintaining speeds closer to a desired target speed. In addition, we show that our reward design enables a generally smaller choice for the discount factor for value-function-based DRL algorithms, which can lead to faster convergence. This is possible since the reward design merely penalizes the one-step decay of the Lyapunov function. Furthermore, policy training simulations employ an early episode termination method to constrain exploration and add more valuable samples to the DRL training replay memory. Finally, real-world experiments with a quadrotor validate the ability of our method to safely navigate around varying densities of obstacles. The proposed method consistently takes cautious maneuvers near obstacles by slowing down, achieving greater obstacle clearance compared to baseline, although with an increase in mission completion time. |
---|---|
AbstractList | The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL reward design inspired by Lyapunov stability theory for safe robot navigation in the presence of obstacles. The navigation problem is formulated as a state-space control problem with close obstacle locations integrated into the state representation. To ensure safe obstacle avoidance, we introduce a novel reward-shaping strategy utilizing a Lyapunov function that discourages fast movement toward obstacles. Our numerical experiments demonstrate the effectiveness of the reward design strategy compared to baselines in achieving consistent superior learning with higher mission completion rates while maintaining speeds closer to a desired target speed. In addition, we show that our reward design enables a generally smaller choice for the discount factor for value-function-based DRL algorithms, which can lead to faster convergence. This is possible since the reward design merely penalizes the one-step decay of the Lyapunov function. Furthermore, policy training simulations employ an early episode termination method to constrain exploration and add more valuable samples to the DRL training replay memory. Finally, real-world experiments with a quadrotor validate the ability of our method to safely navigate around varying densities of obstacles. The proposed method consistently takes cautious maneuvers near obstacles by slowing down, achieving greater obstacle clearance compared to baseline, although with an increase in mission completion time. |
Author | Ugurlu, Halil Ibrahim Redder, Adrian Kayacan, Erdal |
Author_xml | – sequence: 1 givenname: Halil Ibrahim surname: Ugurlu fullname: Ugurlu, Halil Ibrahim organization: Aarhus University,Department of Electrical and Computer Engineering,Aarhus,Denmark – sequence: 2 givenname: Adrian surname: Redder fullname: Redder, Adrian organization: Paderborn University,Department of Automatic Control,Paderborn,Germany – sequence: 3 givenname: Erdal surname: Kayacan fullname: Kayacan, Erdal organization: Paderborn University,Department of Automatic Control,Paderborn,Germany |
BookMark | eNo1j8tKw0AYRkfQhda-geC8QOpcMpcsJUYbCBaqrlyUSfJPGWj_CZMY6NtrUVcHDpwPvhtyiRGBkHvOVpyz4qGsqzedF0qtBBPq7JjRwlyQZWEKKyVXzAojrslnc3LDF8Y5q3EcQoKePgEMdAsBfUwdHAEn2oBLGHBPfxTdxjZO9NXNYe-mEJEGpJt2nFx3AFrhHFLEczXekivvDiMs_7ggH8_Ve7nOms1LXT42WeDGTplXkknBmPAtA-OdN8pIW4A10ivfe9HZjvd52-rcCyu91jnoovMsl85IZuSC3P3uBgDYDSkcXTrt_j_LbzuwUkU |
ContentType | Conference Proceeding |
DBID | 6IE 6IL CBEJK RIE RIL |
DOI | 10.1109/CIES64955.2025.11007627 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Xplore POP ALL IEEE Xplore All Conference Proceedings IEEE/IET Electronic Library IEEE Proceedings Order Plans (POP All) 1998-Present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE/IET Electronic Library url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
EISBN | 9798331508272 |
EndPage | 8 |
ExternalDocumentID | 11007627 |
Genre | orig-research |
GroupedDBID | 6IE 6IL CBEJK RIE RIL |
ID | FETCH-LOGICAL-i178t-f53032002fb0e7faf757389e873f5fdf2c8c1d4bb64f283f664e69cf043a73073 |
IEDL.DBID | RIE |
IngestDate | Thu May 29 05:57:27 EDT 2025 |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i178t-f53032002fb0e7faf757389e873f5fdf2c8c1d4bb64f283f664e69cf043a73073 |
PageCount | 8 |
ParticipantIDs | ieee_primary_11007627 |
PublicationCentury | 2000 |
PublicationDate | 2025-March-17 |
PublicationDateYYYYMMDD | 2025-03-17 |
PublicationDate_xml | – month: 03 year: 2025 text: 2025-March-17 day: 17 |
PublicationDecade | 2020 |
PublicationTitle | 2025 IEEE Symposium on Computational Intelligence on Engineering/Cyber Physical Systems (CIES) |
PublicationTitleAbbrev | CIES |
PublicationYear | 2025 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
Score | 1.9047948 |
Snippet | The inherent black-box nature of deep reinforcement learning (DRL) poses challenges in ensuring safety constraints. This paper, therefore, introduces a DRL... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 1 |
SubjectTerms | Collision avoidance Convergence Deep reinforcement learning Lyapunov methods Lyapunov theory Navigation Obstacle avoidance Planning Quadrotor navigation Quadrotors Reward design Robots Safety Training |
Title | Lyapunov-Inspired Deep Reinforcement Learning for Robot Navigation in Obstacle Environments |
URI | https://ieeexplore.ieee.org/document/11007627 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1bS8MwFA66J59UnHgnD7626yW3Ps-NKTplOBj4MJr0RIbQDmkH-us9yTpvIPhWQkNLktPzNfm-7xBymWsJDAPaMdWSwGH0QNtCB8ALbfDrx5Uv33Y3FqMpu5nxWStW91oYAPDkMwjdpT_LLyrTuK2ynrM3w-CV22Qb19larNVytuIo6_URQAkE_Bx_-xIebu7-UTfFp43hLhlvHrhmi7yETa1D8_7Li_Hfb7RHul8KPfrwmXv2yRaUB-Tp9i1fNmW1Cq5Ld4AOBb0CWNIJeHtU43cCaeuo-kyxiU4qXdV0nK-800ZV0kVJ7zUiRlxNdPBNBNcl0-HgsT8K2uIJwSKWqg4sT11t9CixOgJpcyu5RHACSqaW28ImRpm4YFoLZhFiWCEYiMzYiKW5dIF_SDplVcIRoSxniGq5yDKIsQNXRqVKsSKRkZUiU8ek60Zmvlz7Y8w3g3LyR_sp2XET5JhcsTwjnfq1gXNM7bW-8FP6AaTopYs |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1dS8MwFA06H_RJxYnf5sHXdm2Xrz7PjU23KmODgQ-jSW9kCO2QdqC_3iTr_ALBtxJaGpLe3pPknHMRukklB2IC2jLVIs9idE_qTHpAM6nM348KV75tlLD-lNzN6KwWqzstDAA48hn49tKd5WeFquxWWcvam5ng5dtoxyR-QtdyrZq1FQZxq2MgFDOQn5qFX0T9zf0_Kqe4xNHbR8nmlWu-yItfldJX77_cGP_dpwPU_NLo4cfP7HOItiA_Qk_Dt3RZ5cXKG-T2CB0yfAuwxGNwBqnK7QXi2lP1GZsmPC5kUeIkXTmvjSLHixw_SIMZzfeEu99kcE007XUnnb5Xl0_wFiEXpadp21ZHDyItA-A61ZxyA09A8LamOtOREirMiJSMaAMyNGMEWKx0QNopt6F_jBp5kcMJwiQlBtdSFscQmgeoUKItBMkiHmjOYnGKmnZk5su1Q8Z8Myhnf7Rfo93-ZDScDwfJ_Tnas5NleV0hv0CN8rWCS5PoS3nlpvcDfxWo2A |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2025+IEEE+Symposium+on+Computational+Intelligence+on+Engineering%2FCyber+Physical+Systems+%28CIES%29&rft.atitle=Lyapunov-Inspired+Deep+Reinforcement+Learning+for+Robot+Navigation+in+Obstacle+Environments&rft.au=Ugurlu%2C+Halil+Ibrahim&rft.au=Redder%2C+Adrian&rft.au=Kayacan%2C+Erdal&rft.date=2025-03-17&rft.pub=IEEE&rft.spage=1&rft.epage=8&rft_id=info:doi/10.1109%2FCIES64955.2025.11007627&rft.externalDocID=11007627 |