资源论文Error Compensated Quantized SGD and its Applications to Large-scale Distributed Optimization

Error Compensated Quantized SGD and its Applications to Large-scale Distributed Optimization

2020-03-19 | |  58 |   33 |   0

Abstract

Large-scale distributed optimization is of great importance in various applications. For dataparallel based distributed learning, the inter-nod gradient communication often becomes the performance bottleneck. In this paper, we propose the error compensated quantized stochastic gradient descent algorithm to improve the training efficiency. Local gradients are quantized to reduce the communication overhead, and accumulated quantization error is utilized to speed up t convergence. Furthermore, we present theoretical analysis on the convergence behaviour, and demonstrate its advantage over competitors. Extensive experiments indicate that our algorithm can compress gradients by a factor of up to two magnitudes without performance degradation.

上一篇:Is Generator Conditioning Causally Related to GAN Performance?

下一篇:Temporal Poisson Square Root Graphical Models

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...