资源论文Monte-Carlo Expectation Maximization for Decentralized POMDPs Feng Wu † Shlomo Zilberstein ‡ Nicholas R. Jennings †

Monte-Carlo Expectation Maximization for Decentralized POMDPs Feng Wu † Shlomo Zilberstein ‡ Nicholas R. Jennings †

2019-11-11 | |  96 |   41 |   0
Abstract We address two signi?cant drawbacks of state-ofthe-art solvers of decentralized POMDPs (DECPOMDPs): the reliance on complete knowledge of the model and limited scalability as the complexity of the domain grows. We extend a recently proposed approach for solving DEC-POMDPs via a reduction to the maximum likelihood problem, which in turn can be solved using EM. We introduce a model-free version of this approach that employs Monte-Carlo EM (MCEM). While a na??ve implementation of MCEM is inadequate in multiagent settings, we introduce several improvements in sampling that produce high-quality results on a variety of DEC-POMDP benchmarks, including large problems with thousands of agents.

上一篇:Endogenous Boolean Games Paolo Turrini

下一篇:Scaling-Up Security Games with Boundedly Rational Adversaries: A Cutting-Plane Approach

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...