Robust Reinforcement Learning Control Using Integral Quadratic Constraints for Recurrent Neural Networks

The applicability of machine learning techniques for feedback control systems is limited by a lack of stability guarantees. Robust control theory offers a framework for analyzing the stability of feedback control loops, but for the integral quadratic constraint (IQC) framework used here, all compone...

Full description

Saved in:
Bibliographic Details
Published inIEEE transactions on neural networks Vol. 18; no. 4; pp. 993 - 1002
Main Authors Anderson, C.W., Young, P.M., Buehner, M.R., Knight, J.N., Bush, K.A., Hittle, D.C.
Format Journal Article
LanguageEnglish
Published United States IEEE 01.07.2007
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The applicability of machine learning techniques for feedback control systems is limited by a lack of stability guarantees. Robust control theory offers a framework for analyzing the stability of feedback control loops, but for the integral quadratic constraint (IQC) framework used here, all components are required to be represented as linear, time-invariant systems plus uncertainties with, for IQCs used here, bounded gain. In this paper, the stability of a control loop including a recurrent neural network (NN) is analyzed by replacing the nonlinear and time-varying components of the NN with IQCs on their gain. As a result, a range of the NN's weights is found within which stability is guaranteed. An algorithm is demonstrated for training the recurrent NN using reinforcement learning and guaranteeing stability while learning.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ObjectType-Article-2
ObjectType-Feature-1
ISSN:1045-9227
1941-0093
DOI:10.1109/TNN.2007.899520