Analysis of Gradient Descent Methods With Nondiminishing Bounded Errors

Ramaswamy, Arunselvan ; Bhatnagar, Shalabh (2018) Analysis of Gradient Descent Methods With Nondiminishing Bounded Errors IEEE Transactions on Automatic Control, 63 (5). pp. 1465-1471. ISSN 0018-9286

Full text not available from this repository.

Official URL: http://doi.org/10.1109/TAC.2017.2744598

Related URL: http://dx.doi.org/10.1109/TAC.2017.2744598

Abstract

The main aim of this paper is to provide an analysis of gradient descent (GD) algorithms with gradient errors that do not necessarily vanish, asymptotically. In particular, sufficient conditions are presented for both stability (almost sure boundedness of the iterates) and convergence of GD with bounded (possibly) nondiminishing gradient errors. In addition to ensuring stability, such an algorithm is shown to converge to a small neighborhood of the minimum set, which depends on the gradient errors. It is worth noting that the main result of this paper can be used to show that GD with asymptotically vanishing errors indeed converges to the minimum set. The results presented herein are not only more general when compared to previous results, but our analysis of GD with errors is new to the literature to the best of our knowledge. Our work extends the contributions of Mangasarian and Solodov, Bertsekas and Tsitsiklis, and Tadiόand Doucet. Using our framework, a simple yet effective implementation of GD using simultaneous perturbation stochastic approximations, with constant sensitivity parameters, is presented. Another important improvement over many previous results is that there are no “additional” restrictions imposed on the step sizes. In machine learning applications where step sizes are related to learning rates, our assumptions, unlike those of other papers, do not affect these learning rates. Finally, we present experimental results to validate our theory.

Item Type:Article
Source:Copyright of this article belongs to Institute of Electrical and Electronics Engineers.
Keywords:Differential Inclusions (DIS); Gradient Descent (GD) Methods; Nondiminishing Errors; Stability And Convergence; Stochastic Approximation Algorithms.
ID Code:116458
Deposited On:12 Apr 2021 05:56
Last Modified:12 Apr 2021 05:56

Repository Staff Only: item control page