资源论文Decentralized Optimization with Edge Sampling

Decentralized Optimization with Edge Sampling

2019-09-29 | |  106 |   48 |   0
Abstract In this paper, we propose a decentralized distributed algorithm with stochastic communication among nodes, building on a sampling method called “edge sampling”. Such a sampling algorithm allows us to avoid the heavy peer-to-peer communication cost when combining neighboring weights on dense networks while still maintains a comparable convergence rate. In particular, we quantitatively analyze its theoretical convergence properties, as well as the optimal sampling rate over the underlying network. When compared with previous methods, our solution is shown to be unbiased, communication-efficient and suffers from lower sampling variances. These theoretical findings are validated by both numerical experiments on the mixing rates of Markov Chains and distributed machine learning problems

上一篇:DBDNet: Learning Bi-directional Dynamics for Early Action Prediction

下一篇:Deep Recurrent Quantization for Generating Sequential Binary Codes

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...