资源论文Interpreting Neural Network Judgments via Minimal, Stable, and Symbolic Corrections

Interpreting Neural Network Judgments via Minimal, Stable, and Symbolic Corrections

2020-02-18 | |  56 |   39 |   0

Abstract 

We present a new algorithm to generate minimal, stable, and symbolic corrections to an input that will cause a neural network with ReLU activations to change its output. We argue that such a correction is a useful way to provide feedback to a user when the network’s output is different from a desired output. Our algorithm generates such a correction by solving a series of linear constraint satisfaction problems. The technique is evaluated on three neural network models: one predicting whether an applicant will pay a mortgage, one predicting whether a first-order theorem can be proved efficiently by a solver using certain heuristics, and the final one judging whether a drawing is an accurate rendition of a canonical drawing of a cat.

上一篇:Designing by Training: Acceleration Neural Network for Fast High-Dimensional Convolution

下一篇:ChannelNets: Compact and Efficient Convolutional Neural Networks via Channel-Wise Convolutions

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...