资源论文Decoupling Gradient-Like Learning Rules from Representations

Decoupling Gradient-Like Learning Rules from Representations

2020-03-16 | |  62 |   32 |   0

Abstract

In machine learning, learning often corresponds to changing the parameters of a parameterized function. A learning rule is an algorithm or mathematical expression that specifies precisely how the parameters should be changed. When creating a machine learning system, we must make two decisions: what representation should be used (i.e., what parameterized function should be used) and what learning rule should be used to search through the resulting set of representable functions. In this paper we focus on gradient-like learning rules, wherein these two decisions are coupled in a subtle (and often unintentional) way. That is, using the same learning rule with two dif ferent representations that can represent the same sets of functions can result in two different outcomes. After arguing that this coupling is undesir able, particularly when using neural networks, we present a method for partially decoupling these two decisions for a broad class of gradient-like learning rules that span unsupervised learning, reinforcement learning, and supervised learning.

上一篇:PDE-Net: Learning PDEs from Data

下一篇:Essentially No Barriers in Neural Network Energy Landscape

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...