资源论文LEARNING EFFICIENT PARAMETER SERVERS YNCHRONIZATION POLICIES FOR DISTRIBUTED SGD

LEARNING EFFICIENT PARAMETER SERVERS YNCHRONIZATION POLICIES FOR DISTRIBUTED SGD

2020-01-02 | |  128 |   42 |   0

Abstract

We apply a reinforcement learning (RL) based approach to learning optimal synchronization policies used for Parameter Server-based distributed training of machine learning models with Stochastic Gradient Descent (SGD). Utilizing a formal synchronization policy description in the PS-setting, we are able to derive a suitable and compact description of states and actions, allowing us to efficiently use the standard off-the-shelf deep Q-learning algorithm. As a result, we are able to learn synchronization policies which generalize to different cluster environments, different training datasets and small model variations and (most importantly) lead to considerable decreases in training time when compared to standard policies such as bulk synchronous parallel (BSP), asynchronous parallel (ASP), or stale synchronous parallel (SSP). To support our claims we present extensive numerical results obtained from experiments performed in both simulated and real cluster environments. In our experiments training time is reduced by 44% on average in simulated and 28% in real cluster environments, respectively. Moreover, the learned policies generalize to multiple unseen circumstances.

上一篇:UNDERSTANDING THE LIMITATIONS OF CONDITIONALG ENERATIVE MODELS

下一篇:ONCE FOR ALL :T RAIN ONE NETWORK ANDS PECIALIZE IT FOR EFFICIENT DEPLOYMENT

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...