资源论文Optimally Solving Dec-POMDPs as Continuous-State MDPs Jilles Steeve Dibangoye Christopher Amato Olivier Buffet and Franc?ois Charpillet

Optimally Solving Dec-POMDPs as Continuous-State MDPs Jilles Steeve Dibangoye Christopher Amato Olivier Buffet and Franc?ois Charpillet

2019-11-08 | |  71 |   42 |   0
Abstract Optimally solving decentralized partially observable Markov decision processes (Dec-POMDPs) is a hard combinatorial problem. Current algorithms search through the space of full histories for each agent. Because of the doubly exponential growth in the number of policies in this space as the planning horizon increases, these methods quickly become intractable. However, in real world problems, computing policies over the full history space is often unnecessary. True histories experienced by the agents often lie near a structured, low-dimensional manifold embedded into the history space. We show that by transforming a Dec-POMDP into a continuous-state MDP, we are able to find and exploit these low-dimensional representations. Using this novel transformation, we can then apply powerful techniques for solving POMDPs and continuous-state MDPs. By combining a general search algorithm and dimension reduction based on feature selection, we introduce a novel approach to optimally solve problems with significantly longer planning horizons than previous methods.

上一篇:Intention-Aware Routing to Minimise Delays at Electric Vehicle Charging Stations?

下一篇:Elicitation and Approximately Stable Matching with Partial Preferences Joanna Drummond and Craig Boutilier

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...