资源论文Interpretable Adversarial Perturbation in Input Embedding Space for Text

Interpretable Adversarial Perturbation in Input Embedding Space for Text

2019-11-07 | |  50 |   37 |   0
Abstract Following great success in the image processing field, the idea of adversarial training has been applied to tasks in the natural language processing (NLP) field. One promising approach directly applies adversarial training developed in the image processing field to the input word embedding space instead of the discrete input space of texts. However, this approach abandons such interpretability as generating adversarial texts to significantly improve the performance of NLP tasks. This paper restores interpretability to such methods by restricting the directions of perturbations toward the existing words in the input embedding space. As a result, we can straightforwardly reconstruct each input with perturbations to an actual text by considering the perturbations to be the replacement of words in the sentence while maintaining or even improving the task performance1 .

上一篇:A Hierarchical End-to-End Model for Jointly Improving Text Summarization and Sentiment Classification

下一篇:Functional Partitioning of Ontologies for Natural Language Query Completion in Question Answering Systems

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...