资源论文TRANSFERABLE PERTURBATIONS OF DEEP FEATURE DISTRIBUTIONS

TRANSFERABLE PERTURBATIONS OF DEEP FEATURE DISTRIBUTIONS

2020-01-02 | |  59 |   31 |   0

Abstract Almost all current adversarial attacks of CNN classifiers rely on information derived from the output layer of the network. This work presents a new adversarial attack based on the modeling and exploitation of class-wise and layer-wise deep feature distributions. We achieve state-of-the-art targeted blackbox transfer-based attack results for undefended ImageNet models. Further, we place a priority on explainability and interpretability of the attacking process. Our methodology affords an analysis of how adversarial attacks change the intermediate feature distributions of CNNs, as well as a measure of layer-wise and class-wise feature distributional separability/entanglement. We also conceptualize a transition from task/data-specific to model-specific features within a CNN architecture that directly impacts the transferability of adversarial examples.

上一篇:DEEP GRAPH MATCHING CONSENSUS

下一篇:GOING BEYOND TOKEN -LEVEL PRE -TRAINING FORE MBEDDING -BASED LARGE -SCALE RETRIEVAL

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...