Large-Scale Machine Learning with Stochastic Gradient Descent
Léon Bottou ()
Additional contact information
Léon Bottou: NEC Labs America
A chapter in Proceedings of COMPSTAT'2010, 2010, pp 177-186 from Springer
Abstract:
Abstract During the last decade, the data sizes have grown faster than the speed of processors. In this context, the capabilities of statistical machine learning methods is limited by the computing time rather than the sample size. A more precise analysis uncovers qualitatively different tradeoffs for the case of small-scale and large-scale learning problems. The large-scale case involves the computational complexity of the underlying optimization algorithm in non-trivial ways. Unlikely optimization algorithms such as stochastic gradient descent show amazing performance for large-scale problems. In particular, second order stochastic gradient and averaged stochastic gradient are asymptotically efficient after a single pass on the training set.
Keywords: stochastic gradient descent; online learning; efficiency (search for similar items in EconPapers)
Date: 2010
References: Add references at CitEc
Citations:
There are no downloads for this item, see the EconPapers FAQ for hints about obtaining it.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:spr:sprchp:978-3-7908-2604-3_16
Ordering information: This item can be ordered from
http://www.springer.com/9783790826043
DOI: 10.1007/978-3-7908-2604-3_16
Access Statistics for this chapter
More chapters in Springer Books from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().