资源论文EXPLORATORY NOT EXPLANATORY:C OUNTERFACTUAL ANALYSIS OF SALIENCY MAPSFOR DEEP RL

EXPLORATORY NOT EXPLANATORY:C OUNTERFACTUAL ANALYSIS OF SALIENCY MAPSFOR DEEP RL

2020-01-02 | |  52 |   38 |   0

Abstract

Saliency maps have been used to support explanations of deep reinforcement learning (RL) agent behavior over temporally extended sequences. However, their use in the community indicates that the explanations derived from saliency maps are often unfalsifiable and can be highly subjective. We introduce an empirical approach grounded in counterfactual reasoning to test the hypotheses generated from saliency maps and assess the degree to which saliency maps represent semantics of RL environments. We evaluate three types of saliency maps using Atari games, a common benchmark for deep RL. Our results show the extent to which existing claims about Atari games can be evaluated and suggest that saliency maps are an exploratory tool not an explanatory tool.

上一篇:MONOTONIC MULTIHEAD ATTENTION

下一篇:DOUBLE NEURAL COUNTERFACTUAL REGRET MINI-MIZATION

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...