资源论文Adaptive Methods for Nonconvex Optimization

Adaptive Methods for Nonconvex Optimization

2020-02-14 | |  65 |   47 |   0

Abstract

 Adaptive gradient methods that rely on scaling gradients down by the square root of exponential moving averages of past squared gradients, such RMS PROP, A DAM, A DADELTA have found wide application in optimizing the nonconvex problems that arise in deep learning. However, it has been recently demonstrated that such methods can fail to converge even in simple convex optimization settings. In this work, we provide a new analysis of such methods applied to nonconvex stochastic optimization problems, characterizing the effect of increasing minibatch size. Our analysis shows that under this scenario such methods do converge to stationarity up to the statistical limit of variance in the stochastic gradients (scaled by a constant factor). In particular, our result implies that increasing minibatch sizes enables convergence, thus providing a way to circumvent the nonconvergence issues. Furthermore, we provide a new adaptive optimization algorithm, YOGI, which controls the increase in effective learning rate, leading to even better performance with similar theoretical guarantees on convergence. Extensive experiments show that YOGI with very little hyperparameter tuning outperforms methods such as A DAM in several challenging machine learning tasks.

上一篇:A theory on the absence of spurious solutions for nonconvex and nonsmooth optimization

下一篇:Dendritic cortical microcircuits approximate the backpropagation algorithm

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...