资源论文A Lower Bound for the Optimization of Finite Sums

A Lower Bound for the Optimization of Finite Sums

2020-03-05 | |  54 |   30 |   0

Abstract

This paper presents a lower bound for optimizing a finite sum of n functions, where each function is L-smooth and the sum is µ-strongly convex. We show that no algorithm can reach an error ε in minimizing all p functions from this class in fewer than 图片.png iterations, where κ = L/µ is a surrogate condition number. We then compare this lower bound to upper bounds for recently developed methods specializing to this setting. When the functions involved in this sum are not arbitrary, but based on i.i.d. random data, then we further contrast these complexity results with those for optimal first-order methods to directly optimize the sum. The conclusion we draw is that a lot of caution is necessary for an accurate comparison, and identify machine learning scenarios where the new methods help computationally.

上一篇:Convex Formulation for Learning from Positive and Unlabeled Data

下一篇:Stochastic Optimization with Importance Sampling for Regularized Loss Minimization

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...