资源论文Monte-Carlo Exploration for Deterministic Planning

Monte-Carlo Exploration for Deterministic Planning

2019-11-15 | |  62 |   39 |   0

Abstract Search methods based on Monte-Carlo simulation have recently led to breakthrough performance improvements in diffificult game-playing domains such as Go and General Game Playing. Monte-Carlo Random Walk (MRW) planning applies MonteCarlo ideas to deterministic classical planning. In the forward chaining planner ARVAND, MonteCarlo random walks are used to explore the local neighborhood of a search state for action selection. In contrast to the stochastic local search approach used in the recent planner Identidem, random walks yield a larger and unbiased sample of the search neighborhood, and require state evaluations only at the endpoints of each walk. On IPC-4 competition problems, the performance of ARVAND is competitive with state of the art systems

上一篇:A Distributed Control Loop for Autonomous Recovery in a Multi-Agent Plan

下一篇:Planning with Partial Preference Models

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...