资源论文Sparse Variational Inference for Generalized Gaussian Process Models

Sparse Variational Inference for Generalized Gaussian Process Models

2020-03-05 | |  62 |   46 |   0

Abstract

Gaussian processes (GP) provide an attractive machine learning model due to their nonparametric form, their flexibility to capture many types of observation data, and their generic inference procedures. Sparse GP inference algorithms address the cubic complexity of GPs by focusing on a small set of pseudo-samples. To date, such approaches have focused on the simple case of Gaussian observation likelihoods. This paper develops a variational sparse solution for GPs under general likelihoods by providing a new characterization of the gradients required for inference in terms of individual observation likelihood terms. In addition, we propose a simple new approach for optimizing the sparse variational approximation using a fixed point computation. We demonstrate experimentally that the fixed point operator acts as a contraction in many cases and therefore leads to fast convergence. An experimental evaluation for count regression, classification, and or dinal regression illustrates the generality and advantages of the new approach.

上一篇:Stochastic Primal-Dual Coordinate Method for Regularized Empirical Risk Minimization

下一篇:Learning Fast-Mixing Models for Structured Prediction

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...