中文版 | English
Title

Generalization performance of multi-pass stochastic gradient descent with convex loss functions

Author
Publication Years
2021
Source Title
ISSN
1532-4435
EISSN
1533-7928
Volume22
Abstract
Stochastic gradient descent (SGD) has become the method of choice to tackle large-scale datasets due to its low computational cost and good practical performance. Learning rate analysis, either capacity-independent or capacity-dependent, provides a unifying viewpoint to study the computational and statistical properties of SGD, as well as the implicit regularization by tuning the number of passes. Existing capacity-independent learning rates require a nontrivial bounded subgradient assumption and a smoothness assumption to be optimal. Furthermore, existing capacity-dependent learning rates are only established for the specific least squares loss with a special structure. In this paper, we provide both optimal capacity-independent and capacity-dependent learning rates for SGD with general convex loss functions. Our results require neither bounded subgradient assumptions nor smoothness assumptions, and are stated with high probability. We achieve this improvement by a refined estimate on the norm of SGD iterates based on a careful martingale analysis and concentration inequalities on empirical processes. © 2021 Yunwen Lei, Ting Hu and Ke Tang.
Keywords
URL[Source Record]
Language
English
SUSTech Authorship
Others
ESI Research Field
COMPUTER SCIENCE
Scopus EID
2-s2.0-85105877079
Data Source
Scopus
Document TypeJournal Article
Identifierhttp://kc.sustech.edu.cn/handle/2SGJ60CL/402795
DepartmentResearch Institute of Trustworthy Autonomous Systems
工学院_计算机科学与工程系
Affiliation
1.School of Computer Science,University of Birmingham,Birmingham,B152TT,United Kingdom
2.School of Mathematics and Statistics,Wuhan University,Wuhan,430072,China
3.Research Institute of Trustworthy Autonomous Systems,Department of Computer Science and Engineering,Southern University of Science and Technology,Shenzhen,518055,China
Recommended Citation
GB/T 7714
Lei,Yunwen,Hu,Ting,Tang,Ke. Generalization performance of multi-pass stochastic gradient descent with convex loss functions[J]. JOURNAL OF MACHINE LEARNING RESEARCH,2021,22.
APA
Lei,Yunwen,Hu,Ting,&Tang,Ke.(2021).Generalization performance of multi-pass stochastic gradient descent with convex loss functions.JOURNAL OF MACHINE LEARNING RESEARCH,22.
MLA
Lei,Yunwen,et al."Generalization performance of multi-pass stochastic gradient descent with convex loss functions".JOURNAL OF MACHINE LEARNING RESEARCH 22(2021).
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Export to Excel
Export to Csv
Altmetrics Score
Google Scholar
Similar articles in Google Scholar
[Lei,Yunwen]'s Articles
[Hu,Ting]'s Articles
[Tang,Ke]'s Articles
Baidu Scholar
Similar articles in Baidu Scholar
[Lei,Yunwen]'s Articles
[Hu,Ting]'s Articles
[Tang,Ke]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Lei,Yunwen]'s Articles
[Hu,Ting]'s Articles
[Tang,Ke]'s Articles
Terms of Use
No data!
Social Bookmark/Share
No comment.

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.