资源论文Sparse Reinforcement Learning via Convex Optimization

Sparse Reinforcement Learning via Convex Optimization

2020-03-04 | |  103 |   50 |   0

Abstract

We propose two new algorithms for the sparse reinforcement learning problem based on different formulations. The first algorithm is an off-line method based on the alternating direction method of multipliers for solving a constrained formulation that explicitly controls the projected Bellman residual. The second algorithm is an online stochastic approximation algorithm that employs the regularized dual averaging technique, using the Lagrangian formulation. The convergence of both algorithms are established. We demonstrate the performance of these algorithms through several classical examples.

上一篇:PAC-inspired Option Discovery in Lifelong Reinforcement Learning

下一篇:Standardized Mutual Information for Clustering Comparisons: One Step Further in Adjustment for Chance

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...