资源论文Nonparametric Bayesian Inverse Reinforcement Learning for Multiple Reward Functionss

Nonparametric Bayesian Inverse Reinforcement Learning for Multiple Reward Functionss

2020-01-16 | |  102 |   48 |   0

Abstract

We present a nonparametric Bayesian approach to inverse reinforcement learning (IRL) for multiple reward functions. Most previous IRL algorithms assume that the behaviour data is obtained from an agent who is optimizing a single reward function, but this assumption is hard to guarantee in practice. Our approach is based on integrating the Dirichlet process mixture model into Bayesian IRL. We provide an efficient Metropolis-Hastings sampling algorithm utilizing the gradient of the posterior to estimate the underlying reward functions, and demonstrate that our approach outperforms previous ones via experiments on a number of problem domains.

上一篇:A P300 BCI for the Masses: Prior Information Enables Instant Unsupervised Spelling

下一篇:Efficient Bayes-Adaptive Reinforcement Learning using Sample-Based Search

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...