资源论文Some Considerations on Learning to Explore via Meta-Reinforcement Learning

Some Considerations on Learning to Explore via Meta-Reinforcement Learning

2020-02-13 | |  81 |   50 |   0

Abstract

We interpret meta-reinforcement learning as the problem of learning how to quickly find a good sampling distribution in a new environment. This interpretation leads to the development of two new meta-reinforcement learning algorithms: E-MAML and E-图片.png . Results are presented on a new environment we call ‘Krazy World’: a difficult high-dimensional gridworld which is designed to highlight the importance of correctly differentiating through sampling distributions in meta-reinforcement learning. Further results are presented on a set of maze environments. We show E-MAML and E-图片.png deliver better performance than baseline algorithms on both tasks.

上一篇:Constrained Cross-Entropy Method for Safe Reinforcement Learning

下一篇:A Lyapunov-based Approach to Safe Reinforcement Learning

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...