资源论文Knowledge Transfer with Jacobian Matching

Knowledge Transfer with Jacobian Matching

2020-03-19 | |  80 |   42 |   0

Abstract

Classical distillation methods transfer representations from a “teacher” neural network to a “student” network by matching their output activations. Recent methods also match their Jacobians, or the gradient of output activations with the inpu However, this involves making some ad hoc decisions, in particular, the choice of the loss functi In this paper, we first establish an equivalence between Jacobian matching and distillation with input noise, from which we derive appropriate loss functions for Jacobian matching. We then rely on this analysis to apply Jacobian matching to transfer learning by establishing equivalence of a recent transfer learning procedure to distilla tion. We then show experimentally on standard image datasets that Jacobian-based penalties improve distillation, robustness to noisy inputs, and transfer learning.

上一篇:Neural Inverse Rendering for General Reflectance Photometric Stereo

下一篇:Learning Localized Spatio-Temporal Models From Streaming Data

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...