资源论文Efficient Communication in Multi-Agent Reinforcement Learning via Variance Based Control

Efficient Communication in Multi-Agent Reinforcement Learning via Variance Based Control

2020-02-26 | |  87 |   55 |   0

Abstract

Multi-agent reinforcement learning (MARL) has recently received considerable attention due to its applicability to a wide range of real-world applications. However, achieving efficient communication among agents has always been an overarching problem in MARL. In this work, we propose Variance Based Control (VBC), a simple yet efficient technique to improve communication efficiency in MARL. By limiting the variance of the exchanged messages between agents during the training phase, the noisy component in the messages can be eliminated effectively, while the useful part can be preserved and utilized by the agents for better performance. Our evaluation using multiple MARL benchmarks indicates that our method achieves 2 - 10× lower in communication overhead than state-of-the-art MARL algorithms, while allowing agents to achieve better overall performance.

上一篇:Propagating Uncertainty in Reinforcement Learning via Wasserstein Barycenters

下一篇:Distributional Reward Decomposition for Reinforcement Learning

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Joint Pose and Ex...

    Facial expression recognition (FER) is a challe...