资源论文Blazing the trails before beating the path: Sample-efficient Monte-Carlo planning

Blazing the trails before beating the path: Sample-efficient Monte-Carlo planning

2020-02-05 | |  56 |   44 |   0

Abstract 

You are a robot and you live in a Markov decision process (MDP) with a finite or an infinite number of transitions from state-action to next states. You got brains and so you plan before you act. Luckily, your roboparents equipped you with a generative model to do some Monte-Carlo planning. The world is waiting for you and you have no time to waste. You want your planning to be efficient. Sample-efficient. Indeed, you want to exploit the possible structure of the MDP by exploring only a subset of states reachable by following near-optimal policies. You want guarantees on sample complexity that depend on a measure of the quantity of near-optimal states. You want something, that is an extension of Monte-Carlo sampling (for estimating an expectation) to problems that alternate maximization (over actions) and expectation (over next states). But you do not want to stop with exponential running time, you want something simple to implement and computationally efficient. You want it all and you want it now. You want TrailBlazer.

上一篇:Bi-Objective Online Matching and Submodular Allocations

下一篇:Variance Reduction in Stochastic Gradient Langevin Dynamics

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...