资源论文JACOBIAN ADVERSARIALLY REGULARIZEDN ETWORKS FOR ROBUSTNESS

JACOBIAN ADVERSARIALLY REGULARIZEDN ETWORKS FOR ROBUSTNESS

2020-01-02 | |  56 |   42 |   0

Abstract

Adversarial examples are crafted with imperceptible perturbations with the intent to fool neural networks. Against such attacks, adversarial training and its variants stand as the strongest defense to date. Previous studies have pointed out that robust models that have undergone adversarial training tend to produce more salient and interpretable Jacobian matrices than their non-robust counterparts. A natural question is whether a model trained with an objective to produce salient Jacobian can result in better robustness. This paper answers this question with affirmative empirical results. We propose Jacobian Adversarially Regularized Networks (JARN) as a method to optimize the saliency of a classifier’s Jacobian by adversarially regularizing the model’s Jacobian to resemble natural training images. Image classifiers trained with JARN show improved robust accuracy compared to standard models on the MNIST, SVHN and CIFAR-10 datasets, uncovering a new angle to boost robustness without using adversarial training examples.

上一篇:GLOBAL RELATIONAL MODELS OF SOURCE CODE

下一篇:EXPLAIN YOUR MOVE :U NDERSTANDING AGENT AC -TIONS USING FOCUSED FEATURE SALIENCY

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...