资源论文CYCLICAL STOCHASTIC GRADIENT MCMC FORBAYESIAN DEEP LEARNING

CYCLICAL STOCHASTIC GRADIENT MCMC FORBAYESIAN DEEP LEARNING

2020-01-02 | |  62 |   36 |   0

Abstract

The posteriors over neural network weights are high dimensional and multimodal. Each mode typically characterizes a meaningfully different representation of the data. We develop Cyclical Stochastic Gradient MCMC (SG-MCMC) to automatically explore such distributions. In particular, we propose a cyclical stepsize schedule, where larger steps discover new modes, and smaller steps characterize each mode. We prove non-asymptotic convergence theory of our proposed algorithm. Moreover, we provide extensive experimental results, including ImageNet, to demonstrate the effectiveness of cyclical SG-MCMC in learning complex multimodal distributions, especially for fully Bayesian inference with modern deep neural networks.

上一篇:Q- LEARNING WITH UCB EXPLORATION IS SAMPLEE FFICIENT FOR INFINITE -H ORIZON MDP

下一篇:BATCH -S HAPING FOR LEARNINGC ONDITIONAL CHANNEL GATED NETWORKS

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...