跳到主要导航 跳到搜索 跳到主要内容

Uniform-in-time weak error analysis for stochastic gradient descent algorithms via diffusion approximation

  • Yuanyuan Feng
  • , Tingran Gao
  • , Lei Li*
  • , Jian Guo Liu
  • , Yulong Lu
  • *此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Diffusion approximation provides weak approximation for stochastic gradient descent algorithms in a finite time horizon. In this paper, we introduce new tools motivated by the backward error analysis of numerical stochastic differential equations into the theoretical framework of diffusion approximation, extending the validity of the weak approximation from finite to infinite time horizon. The new techniques developed in this paper enable us to characterize the asymptotic behavior of constant-step-size SGD algorithms near a local minimum around which the objective functions are locally strongly convex, a goal previously unreachable within the diffusion approximation framework. Our analysis builds upon a truncated formal power expansion of the solution of a Kolmogorov equation arising from diffusion approximation, where the main technical ingredient is uniform-in-time bounds controlling the long-term behavior of the expansion coefficient functions near the local minimum. We expect these new techniques to bring new understanding of the behaviors of SGD near local minimum and greatly expand the range of applicability of diffusion approximation to cover wider and deeper aspects of stochastic optimization algorithms in data science.

源语言英语
页(从-至)163-188
页数26
期刊Communications in Mathematical Sciences
18
1
DOI
出版状态已出版 - 2020
已对外发布

指纹

探究 'Uniform-in-time weak error analysis for stochastic gradient descent algorithms via diffusion approximation' 的科研主题。它们共同构成独一无二的指纹。

引用此