Borkar, V. S. (2005) An actor-critic algorithm for constrained markov decision processes Systems & Control Letters, 54 (3). pp. 207-213. ISSN 0167-6911
Full text not available from this repository.
Official URL: http://linkinghub.elsevier.com/retrieve/pii/S01676...
Related URL: http://dx.doi.org/10.1016/j.sysconle.2004.08.007
Abstract
An actor-critic type reinforcement learning algorithm is proposed and analyzed for constrained controlled Markov decision processes. The analysis uses multiscale stochastic approximation theory and the envelope theorem' of mathematical economics.
Item Type: | Article |
---|---|
Source: | Copyright of this article belongs to Elsevier Science. |
Keywords: | Actor-critic Algorithms; Reinforcement Learning; Constrained Markov Decision Processes; Stochastic Approximation; Envelope Theorem |
ID Code: | 5285 |
Deposited On: | 18 Oct 2010 08:32 |
Last Modified: | 20 May 2011 08:53 |
Repository Staff Only: item control page