资源论文Scaling Up Optimal Heuristic Search in Dec-POMDPs via Incremental Expansion

Scaling Up Optimal Heuristic Search in Dec-POMDPs via Incremental Expansion

2019-11-12 | |  113 |   85 |   0
Abstract Planning under uncertainty for multiagent systems can be formalized as a decentralized partially observable Markov decision process. We advance the state of the art for optimal solution of this model, building on the Multiagent A* heuristic search method. A key insight is that we can avoid the full expansion of a search node that generates a number of children that is doubly exponential in the node’s depth. Instead, we incrementally expand the children only when a next child might have the highest heuristic value. We target a subsequent bottleneck by introducing a more memory-ef?cient representation for our heuristic functions. Proof is given that the resulting algorithm is correct and experiments demonstrate a signi?cant speedup over the state of the art, allowing for optimal solutions over longer horizons for many benchmark problems.

上一篇:Replanning in Domains with Partial Information and Sensing Actions

下一篇:On the Effectiveness of CNF and DNF Representations in Contingent Planning

用户评价
全部评价

热门资源

  • Regularizing RNNs...

    Recently, caption generation with an encoder-de...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Deep Cross-media ...

    Cross-media retrieval is a research hotspot in ...

  • Joint Pose and Ex...

    Facial expression recognition (FER) is a challe...

  • Supervised Descen...

    Many computer vision problems (e.