Premium
A reinforcement learning‐based scheme for direct adaptive optimal control of linear stochastic systems
Author(s) -
Wong Wee Chin,
Lee Jay H.
Publication year - 2010
Publication title -
optimal control applications and methods
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.458
H-Index - 44
eISSN - 1099-1514
pISSN - 0143-2087
DOI - 10.1002/oca.915
Subject(s) - reinforcement learning , generalization , mathematical optimization , convergence (economics) , computer science , adaptive control , controller (irrigation) , stochastic control , nonlinear system , optimal control , linear system , adaptive learning , control theory (sociology) , chemostat , scheme (mathematics) , mathematics , control (management) , artificial intelligence , mathematical analysis , physics , genetics , quantum mechanics , bacteria , agronomy , economics , biology , economic growth
Abstract Reinforcement learning where decision‐making agents learn optimal policies through environmental interactions is an attractive paradigm for model‐free, adaptive controller design. However, results for systems with continuous state and action variables are rare. In this paper, we present convergence results for optimal linear quadratic control of discrete‐time linear stochastic systems. This work can be viewed as a generalization of a previous work on deterministic linear systems. Key differences between the algorithms for deterministic and stochastic systems are highlighted through examples. The usefulness of the algorithm is demonstrated through a nonlinear chemostat bioreactor case study. Copyright © 2009 John Wiley & Sons, Ltd.