Trajectory tracking control for Ackerman vehicle based on improved reward function
This article focuses on an online reinforcement learning algorithm (ORRL) for trajectory tracking control of a class of nonlinear systems. Based on the Q-learning framework, the trajectory tracking problem can be solved under the complex paths by designing an optimal reward function integrating mult...
Saved in:
Published in | 2024 43rd Chinese Control Conference (CCC) pp. 519 - 524 |
---|---|
Main Authors | , , , |
Format | Conference Proceeding |
Language | English |
Published |
Technical Committee on Control Theory, Chinese Association of Automation
28.07.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | This article focuses on an online reinforcement learning algorithm (ORRL) for trajectory tracking control of a class of nonlinear systems. Based on the Q-learning framework, the trajectory tracking problem can be solved under the complex paths by designing an optimal reward function integrating multiple sub objectives. The number of iterations and the complexity of the algorithm are reduced by the multi-objective reward mechanism. At the same time, the information of sub-objectives is fully considered to provide richer decision-making information, such that the control process is smoother. Moreover, the short-term action decision and long-term control effect can be balanced by the ORRL. The real vehicle results are given to compared the MPC, Stanley, LQR, standard reinforcement learning algorithm and the ORRL method, and show the higher trajectory tracking accuracy, smoothness in complex trajectory scenes and better adaptive ability. |
---|---|
ISSN: | 1934-1768 |
DOI: | 10.23919/CCC63176.2024.10662434 |