资源论文IMPROVING NEURAL LANGUAGE GENERATION WITHS PECTRUM CONTROL

IMPROVING NEURAL LANGUAGE GENERATION WITHS PECTRUM CONTROL

2019-12-31 | |  71 |   44 |   0

Abstract

Recent Transformer-based models such as Transformer-XL and BERT have achieved huge success on various natural language processing tasks. However, contextualized embeddings at the output layer of these powerful models tend to degenerate and occupy an anisotropic cone in the vector space, which is called the representation degeneration problem. In this paper, we propose a novel spectrum control approach to address this degeneration problem. The core idea of our method is to directly guide the spectra training of the output embedding matrix with a slow-decaying singular value prior distribution through a reparameterization framework. We show that our proposed method encourages isotropy of the learned word representations while maintains the modeling power of these contextual neural models. We further provide a theoretical analysis and insight on the benefit of modeling singular value distribution. We demonstrate that our spectrum control method outperforms the state-of-the-art Transformer-XL modeling for language model, and various Transformer-based models for machine translation, on common benchmark datasets for these tasks.

上一篇:GENERALIZATION THROUGH MEMORIZATION :N EAREST NEIGHBOR LANGUAGE MODELS

下一篇:CROSS -L INGUAL ABILITY OF MULTILINGUAL BERT:A NE MPIRICAL STUDY

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...