Adaptive actor-critic control of robots with integral invariant manifold
The actor-critic scheme stands for a powerful algorithm to design controllers for linear and non-linear systems subject to changing or highly uncertain dynamics. In particular, the actor-critic scheme that has succeeded is typically based on two neural network stages in a hierarchical architecture w...
Saved in:
Published in | 2021 IEEE CHILEAN Conference on Electrical, Electronics Engineering, Information and Communication Technologies (CHILECON) pp. 1 - 6 |
---|---|
Main Authors | , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
06.12.2021
|
Subjects | |
Online Access | Get full text |
DOI | 10.1109/CHILECON54041.2021.9703056 |
Cover
Loading…
Summary: | The actor-critic scheme stands for a powerful algorithm to design controllers for linear and non-linear systems subject to changing or highly uncertain dynamics. In particular, the actor-critic scheme that has succeeded is typically based on two neural network stages in a hierarchical architecture where the critic stage approximates the reward cost function. In contrast, the dynamic of the system is estimated by another neural network in the actor stage. This paper proposes an adaptive actor-critic robot learning on a lower dimension invariant error manifold as part of the Performance Evaluator. The proposed scheme guarantees an envelope of exponential convergence of tracking errors using a modified Lyapunov function, throughout integral sliding mode enforced for all time, where this becomes fundamental to drive also the learning of Reward function. Simulations show a non-linear dynamical robot learning tracking a time-varying trajectory under this Reinforcement Learning scheme. |
---|---|
DOI: | 10.1109/CHILECON54041.2021.9703056 |