资源论文Neural Machine Translation with Key-Value Memory-Augmented Attention

Neural Machine Translation with Key-Value Memory-Augmented Attention

2019-11-07 | |  65 |   51 |   0
Abstract Although attention-based Neural Machine Translation (NMT) has achieved remarkable progress in recent years, it still suffers from issues of repeating and dropping translations. To alleviate these issues, we propose a novel key-value memory-augmented attention model for NMT, called KVM EM ATT. Specifically, we maintain a timely updated keymemory to keep track of attention history and a fixed value-memory to store the representation of source sentence throughout the whole translation process. Via nontrivial transformations and iterative interactions between the two memories, the decoder focuses on more appropriate source word(s) for predicting the next target word at each decoding step, therefore can improve the adequacy of translations. Experimental results on Chinese?English and WMT17 German?English translation tasks demonstrate the superiority of the proposed model.

上一篇:AAR-CNNs: Auto Adaptive Regularized Convolutional Neural Networks

下一篇:Reachability Analysis of Deep Neural Networks with Provable Guarantees

用户评价
全部评价

热门资源

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...