资源论文Bandit Smooth Convex Optimization: Improving the Bias-Variance Tradeoff

Bandit Smooth Convex Optimization: Improving the Bias-Variance Tradeoff

2020-02-04 | |  80 |   48 |   0

Abstract

 Bandit convex optimization is one of the fundamental problems in the field of online learning. The best algorithm for the general bandit convex optimization problem guarantees a regret of image.png while the best known lower bound is image.pngMany attempts have been made to bridge the huge gap between these bounds. A particularly interesting special case of this problem assumes that the loss functions are smooth. In this case, the best known algorithm guarantees a ree image.png We present an efficient algorithm for the bandit smooth convex  optimization problem that guarantees a regret of image.png Our result rules out an image.png lower bound and takes a significant step towards the resolution of this open problem.

上一篇:Learning to Segment Object Candidates

下一篇:Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...