资源论文Bandit Learning with Implicit Feedback

Bandit Learning with Implicit Feedback

2020-02-14 | |  50 |   40 |   0

Abstract

 Implicit feedback, such as user clicks, although abundant in online information service systems, does not provide substantial evidence on users’ evaluation of system’s output. Without proper modeling, such incomplete supervision inevitably misleads model estimation, especially in a bandit learning setting where the feedback is acquired on the fly. In this work, we perform contextual bandit learning with implicit feedback by modeling the feedback as a composition of user result examination and relevance judgment. Since users’ examination behavior is unobserved, we introduce latent variables to model it. We perform Thompson sampling on top of variational Bayesian inference for arm selection and model update. Our upper regret bound analysis of the proposed algorithm proves its feasibility of learning from implicit feedback in a bandit setting; and extensive empirical evaluations on click logs collected from a major MOOC platform further demonstrate its learning effectiveness in practice.

上一篇:Theoretical Linear Convergence of Unfolded ISTA and its Practical Weights and Thresholds

下一篇:Nearly tight sample complexity bounds for learning mixtures of Gaussians via sample compression schemes

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...