资源论文Equivalence Relations in Fully and Partially Observable Markov Decision Processes

Equivalence Relations in Fully and Partially Observable Markov Decision Processes

2019-11-15 | |  67 |   44 |   0
Abstract We explore equivalence relations between states in Markov Decision Processes and Partially Observable Markov Decision Processes. We focus on two different equivalence notions: bisimulation [Givan et al., 2003] and a notion of trace equivalence, under which states are considered equivalent if they generate the same conditional probability distributions over observation sequences (where the conditioning is on action sequences). We show that the relationship between these two equivalence notions changes depending on the amount and nature of the partial observability. We also present an alternate characterization of bisimulation based on trajectory equivalence.

上一篇:Incremental Heuristic Search for Planning with Temporally Extended Goals and Uncontrollable Events

下一篇:Completeness and Optimality Preserving Reduction for Planning

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...