资源论文ADVERSARIAL POLICIES :ATTACKINGD EEP REINFORCEMENT LEARNING

ADVERSARIAL POLICIES :ATTACKINGD EEP REINFORCEMENT LEARNING

2020-01-02 | |  80 |   56 |   0

Abstract

Deep reinforcement learning (RL) policies are known to be vulnerable to adversarial perturbations to their observations, similar to adversarial examples for classifiers. However, an attacker is not usually able to directly modify another agent’s observations. This might lead one to wonder: is it possible to attack an RL agent simply by choosing an adversarial policy acting in a multi-agent environment so as to create natural observations that are adversarial? We demonstrate the existence of adversarial policies in zero-sum games between simulated humanoid robots with proprioceptive observations, against state-of-the-art victims trained via self-play to be robust to opponents. The adversarial policies reliably win against the victims but generate seemingly random and uncoordinated behavior. We find that these policies are more successful in high-dimensional environments, and induce substantially different activations in the victim policy network than when the victim plays against a normal opponent. Fine-tuning protects a victim against a specific adversary, but the attack method can be successfully reapplied to find a new adversarial policy. Videos are available at https://attackingrl.github.io/.

上一篇:BLACK -BOX OFF -POLICY ESTIMATION FORI NFINITE -H ORIZON REINFORCEMENT LEARNING

下一篇:IMPROVING GENERALIZATION IN META REINFORCE -MENT LEARNING USING LEARNED OBJECTIVES

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...