SIAM Review, volume 60, issue 2, pages 223-311

Optimization Methods for Large-Scale Machine Learning

Léon Bottou 1
Frank E. Curtis 2
Jorge Nocedal 3
1
 
FACEBOOK, INC.,
3
 
Industrial Engineering and Management Sciences
Publication typeJournal Article
Publication date2018-05-03
Journal: SIAM Review
scimago Q1
SJR2.900
CiteScore16.9
Impact factor10.8
ISSN00361445, 10957200
Computational Mathematics
Applied Mathematics
Theoretical Computer Science
Abstract
This paper provides a review and commentary on the past, present, and future of numerical optimization algorithms in the context of machine learning applications. Through case studies on text classification and the training of deep neural networks, we discuss how optimization problems arise in machine learning and what makes them challenging. A major theme of our study is that large-scale machine learning represents a distinctive setting in which the stochastic gradient (SG) method has traditionally played a central role while conventional gradient-based nonlinear optimization techniques typically falter. Based on this viewpoint, we present a comprehensive theory of a straightforward, yet versatile SG algorithm, discuss its practical behavior, and highlight opportunities for designing algorithms with improved performance. This leads to a discussion about the next generation of optimization methods for large-scale machine learning, including an investigation of two main streams of research on techniques th...
Found 
Found 

Top-30

Journals

5
10
15
20
25
30
35
40
5
10
15
20
25
30
35
40

Publishers

100
200
300
400
500
600
100
200
300
400
500
600
  • We do not take into account publications without a DOI.
  • Statistics recalculated only for publications connected to researchers, organizations and labs registered on the platform.
  • Statistics recalculated weekly.

Are you a researcher?

Create a profile to get free access to personal recommendations for colleagues and new articles.
Share
Cite this
GOST | RIS | BibTex | MLA
Found error?