资源论文Finito: A Faster, Permutable Incremental Gradient Method for Big Data Problems

Finito: A Faster, Permutable Incremental Gradient Method for Big Data Problems

2020-03-03 | |  61 |   36 |   0

Abstract

Recent advances in optimization theory have shown that smooth strongly convex finite sums can be minimized faster than by treating them as a black box ”batch” problem. In this work we introduce a new method in this class with a theoretical convergence rate four times faster than existing methods, for sums with sufficiently many terms. This method is also amendable to a sampling without replacement scheme that in practice gives further speed-ups. We give empirical results showing state of the art performance.

上一篇:Asymptotically consistent estimation of the number of change points in highly dependent time series

下一篇:Active Transfer Learning under Model Shift

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...