资源论文Byzantine Stochastic Gradient Descent

Byzantine Stochastic Gradient Descent

2020-02-17 | |  44 |   40 |   0

Abstract 

This paper studies the problem of distributed stochastic optimization in an adversarial setting where, out of m machines which allegedly compute stochastic gradients every iteration, an image.png-fraction are Byzantine, and may behave adversarially. Our main result is a variant of stochastic gradient descent (SGD) which finds image.png-approximate minimizers of convex functions in image.png iterations. " m " In contrast, traditional mini-batch SGD needs image.png iterations, but cannot tolerate Byzantine failures. Further, we provide a lower bound showing that, up to logarithmic factors, our algorithm is information-theoretically optimal both in terms of sample complexity and time complexity.

上一篇:On gradient regularizers for MMD GANs

下一篇:Domain-Invariant Projection Learning for Zero-Shot Recognition

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...