Borkar, V. S. ; Meyn, S. P. (2000) The O.D.E. method for convergence of stochastic approximation and reinforcement learning SIAM Journal on Control and Optimization, 38 (2). pp. 447-469. ISSN 0363-0129
|
PDF
- Publisher Version
400kB |
Official URL: http://portal.acm.org/citation.cfm?id=334817.33482...
Related URL: http://dx.doi.org/10.1137/S0363012997331639
Abstract
It is shown here that stability of the stochastic approximation algorithm is implied by the asymptotic stability of the origin for an associated ODE. This in turn implies convergence of the algorithm. Several specific classes of algorithms are considered as applications. It is found that the results provide (i) a simpler derivation of known results for reinforcement learning algorithms; (ii) a proof for the first time that a class of asynchronous stochastic approximation algorithms are convergent without using any a priori assumption of stability; (iii) a proof for the first time that asynchronous adaptive critic and Q-learning algorithms are convergent for the average cost optimal control problem.
Item Type: | Article |
---|---|
Source: | Copyright of this article belongs to Society for Industrial and Applied Mathematics. |
ID Code: | 5333 |
Deposited On: | 18 Oct 2010 08:46 |
Last Modified: | 16 May 2016 15:51 |
Repository Staff Only: item control page