The O.D.E. method for convergence of stochastic approximation and reinforcement learning

Borkar, V. S. ; Meyn, S. P. (2000) The O.D.E. method for convergence of stochastic approximation and reinforcement learning SIAM Journal on Control and Optimization, 38 (2). pp. 447-469. ISSN 0363-0129

[img]
Preview
PDF - Publisher Version
400kB

Official URL: http://portal.acm.org/citation.cfm?id=334817.33482...

Related URL: http://dx.doi.org/10.1137/S0363012997331639

Abstract

It is shown here that stability of the stochastic approximation algorithm is implied by the asymptotic stability of the origin for an associated ODE. This in turn implies convergence of the algorithm. Several specific classes of algorithms are considered as applications. It is found that the results provide (i) a simpler derivation of known results for reinforcement learning algorithms; (ii) a proof for the first time that a class of asynchronous stochastic approximation algorithms are convergent without using any a priori assumption of stability; (iii) a proof for the first time that asynchronous adaptive critic and Q-learning algorithms are convergent for the average cost optimal control problem.

Item Type:Article
Source:Copyright of this article belongs to Society for Industrial and Applied Mathematics.
ID Code:5333
Deposited On:18 Oct 2010 08:46
Last Modified:16 May 2016 15:51

Repository Staff Only: item control page