资源论文PLUG AND PLAY LANGUAGE MODELS :A SIMPLE AP-PROACH TO STEERABLE TEXT GENERATION

PLUG AND PLAY LANGUAGE MODELS :A SIMPLE AP-PROACH TO STEERABLE TEXT GENERATION

2019-12-31 | |  87 |   47 |   0

Abstract
Large transformer-based generative models trained on huge corpora have shown unparalleled language generation ability. While these models are powerful, finegrained control of attributes of the generated language (e.g. gradually switching topic or sentiment) is difficult without modifying the model architecture to allow extra attribute inputs or fine-tuning with attribute-specific data. Both entirely change the original generative function — which, if done poorly, cannot be undone — and also entail the significant cost of retraining. We instead propose the simple Plug and Play Language Model (PPLM) approach for controlled language generation. PPLM consists of plugging in simple attribute classifiers (which may be single layer models or even a bag-of-words), and making updates in the activation space, without changing any model parameters. Such a control scheme provides vast flexibility and allows full recovery of the original generative function. The results demonstrate fine-grained control over a range of topics and sentiment styles, as well as the ability to detoxify generated texts. Our experiments, including human evaluation studies, show that text generated via this control scheme is aligned with desired attributes, while retaining fluency.

上一篇:SELF -A DVERSARIAL LEARNING WITH COMPARATIVED ISCRIMINATION FOR TEXT GENERATION

下一篇:NEURAL TEXT DE GENERATION WITHU NLIKELIHOOD TRAINING

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...