Asymptotic study of stochastic adaptive algorithm in non-convex landscape
Sébastien Gadat and
Ioana Gavra
No 21-1175, TSE Working Papers from Toulouse School of Economics (TSE)
Abstract:
This paper studies some asymptotic properties of adaptive algorithms widely used in optimization and machine learning, and among them Adagrad and Rmsprop, which are involved in most of the blackbox deep learning algorithms. Our setup is the non-convex landscape optimization point of view, we consider a one time scale parametrization and we consider the situation where these algorithms may be used or not with mini-batches. We adopt the point of view of stochastic algorithms and establish the almost sure convergence of these methods when using a decreasing step-size towards the set of critical points of the target function. With a mild extra assumption on the noise, we also obtain the convergence towards the set of minimizers of the function. Along our study, we also obtain a \convergence rate" of the methods, in the vein of the works of [GL13].
Keywords: Stochastic optimization; Stochastic adaptive algorithm; Convergence of random variables (search for similar items in EconPapers)
Date: 2021-01-07
New Economics Papers: this item is included in nep-cmp
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (1)
Downloads: (external link)
https://www.tse-fr.eu/sites/default/files/TSE/docu ... 2021/wp_tse_1175.pdf Full Text (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:tse:wpaper:125116
Access Statistics for this paper
More papers in TSE Working Papers from Toulouse School of Economics (TSE) Contact information at EDIRC.
Bibliographic data for series maintained by ().