资源论文Stochastic Nested Variance Reduction for Nonconvex Optimization

Stochastic Nested Variance Reduction for Nonconvex Optimization

2020-02-13 | |  53 |   35 |   0

Abstract

 We study finite-sum nonconvex optimization problems, where the objective function is an average of n nonconvex functions. We propose a new stochastic gradient descent algorithm based on nested variance reduction. Compared with conventional stochastic variance reduced gradient (SVRG) algorithm that uses two reference points to construct a semi-stochastic gradient with diminishing variance in each iteration, our algorithm uses K + 1 nested reference points to build a semi-stochastic gradient to further reduce its variance in each iteration. For smooth nonconvex functions, the proposed algorithm converges to an ?-approximate first-order stationary point (i.e., image.png within image.png number of stochastic gradient evaluations. This improves the best known gradient complexity of SVRG image.png and that of SCSG image.png . For gradient dominated functions, our algorithm also achieves better gradient complexity than the state-of-the-art algorithms. Thorough experimental results on different nonconvex optimization problems back up our theory.

上一篇:Solving Non-smooth Constrained Programs with Lower Complexity than O(1/ Homotopy Smoothing Approach

下一篇:The challenge of realistic music generation: modelling raw audio at scale

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...