资源论文ME TA -Q -L EA RN IN G

ME TA -Q -L EA RN IN G

2020-01-02 | |  52 |   45 |   0

Abstract

M E TA Q L E A R N I N G Anonymous authors Paper under double-blind review ABSTRACT This paper introduces Meta-Q-Learning (MQL), a new off-policy algorithm for meta-Reinforcement Learning (meta-RL). MQL builds upon three simple ideas. First, we show that Q-learning is competitive with state of the art meta-RL algorithms if given access to a context variable that is a representation of the past trajectory. Second, using a multi-task objective to maximize the average reward across the training tasks is an effective method to meta-train RL policies. Third, past data from the meta-training replay buffer can be recycled to adapt the policy on a new task using off-policy updates. MQL draws upon ideas in propensity estimation to do so and thereby amplifies the amount of available data for adaptation. Experiments on standard continuous-control benchmarks suggest that MQL compares favorably with state of the art meta-RL algorithms.

上一篇:PITFALLS OF IN -D OMAIN UNCERTAINTY ESTIMATIONAND ENSEMBLING IN DEEP LEARNING

下一篇:LEARNING TO LEARN BY ZEROTH -O RDER ORACLE

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...