资源论文DARLA: Improving Zero-Shot Transfer in Reinforcement Learning

DARLA: Improving Zero-Shot Transfer in Reinforcement Learning

2020-03-09 | |  88 |   43 |   0

Abstract

Domain adaptation is an important open problem in deep reinforcement learning (RL). In many scenarios of interest data is hard to obtain, so agents may learn a source policy in a setting where data is readily available, with the hope that it generalises well to the target domain. We propose a new multi-stage RL agent, DARLA (DisentAngled Representation Learning Agent), which learns to see before learning to act. DARLA’s vision is based on learning a disentangled representation of the observed environment. Once DARLA can see, it is able to acquire source policies that are robust to many domain shifts even with no access to the target domain. DARLA significantly outperforms conventional baselines in zero-shot domain adaptation scenarios, an effect that holds across a variety of RL en vironments (Jaco arm, DeepMind Lab) and base RL algorithms (DQN, A3C and EC).

上一篇:State-Frequency Memory Recurrent Neural Networks

下一篇:Neural Taylor Approximations: Convergence and Exploration in Rectifier Networks

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...