资源论文LEARNING TO EXPLORE USINGACTIVE NEURAL MAPPING

LEARNING TO EXPLORE USINGACTIVE NEURAL MAPPING

2020-01-02 | |  58 |   37 |   0

Abstract

This work presents a modular and hierarchical approach to learn policies for exploring 3D environments. Our approach leverages the strengths of both classical and learning-based methods, by using analytical path planners with learned mappers, and global and local policies. Use of learning provides flexibility with respect to input modalities (in mapper), leverages structural regularities of the world (in global policies), and provides robustness to errors in state estimation (in local policies). Such use of learning within each module retains its benefits, while at the same time, hierarchical decomposition and modular training allow us to sidestep the high sample complexities associated with training end-to-end policies. Our experiments in visually and physically realistic simulated 3D environments demonstrate the effectiveness of our approach over past learning and geometry-based approaches.1

上一篇:UNDERSTANDING AND IMPROVING INFORMATIONT RANSFER IN MULTI -TASK LEARNING

下一篇:ADVECTIVE NET: AN EULERIAN -L AGRANGIANF LUIDIC RESERVOIR FOR POINT CLOUD PROCESSING

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...