资源论文Coordinated Exploration in Concurrent Reinforcement Learning

Coordinated Exploration in Concurrent Reinforcement Learning

2020-03-16 | |  71 |   56 |   0

Abstract

We consider a team of reinforcement learning agents that concurrently learn to operate in a com mon environment. We identify three properties – adaptivity, commitment, and diversity – which are necessary for efficient coordinated exploratio and demonstrate that straightforward extensions to single-agent optimistic and posterior sampling approaches fail to satisfy them. As an alternative, we propose seed sampling, which extends posterior sampling in a manner that meets these requirements. Simulation results investigate how per-agent regret decreases as the number of agents grows, establishing substantial advantages of seed sampling over alternative exploration schemes.

上一篇:CRAFTML, an Efficient Clustering-based Random Forest for Extreme Multi-label Learning

下一篇:Matrix Norms in Data Streams: Faster, Multi-Pass and Row-Order

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...