资源论文Online Continuous Submodular Maximization: From Full-Information to Bandit Feedback

Online Continuous Submodular Maximization: From Full-Information to Bandit Feedback

2020-02-20 | |  36 |   36 |   0

Abstract

In this paper, we propose three online algorithms for submodular maximization. The first one, Mono-Frank-Wolfe, reduces the number of per-function gradient evaluations from 图片.png [18] and 图片.png [17] to 1, and achieves a 图片.png-regret bound of 图片.png The second one, Bandit-Frank-Wolfe, is the first bandit algorithm for continuous DR-submodular maximization, which achieves a 图片.png-regret bound of 图片.png. Finally, we extend Bandit-Frank-Wolfe to a bandit algorithm for discrete submodular maximization, Responsive-Frank-Wolfe, which attains a 图片.png-regret bound of 图片.png in the responsive bandit setting.

上一篇:Using Embeddings to Correct for Unobserved Confounding in Networks

下一篇:No-Regret Learning in Unknown Games with Correlated Payoffs

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...