资源论文Unsupervised Neural Text Simplification

Unsupervised Neural Text Simplification

2019-09-24 | |  195 |   103 |   0

Abstract The paper presents a fifirst attempt towards unsupervised neural text simplifification that relies only on unlabeled text corpora. The core framework is composed of a shared encoder and a pair of attentional-decoders, crucially assisted by discrimination-based losses and denoising. The framework is trained using unlabeled text collected from en-Wikipedia dump. Our analysis (both quantitative and qualitative involving human evaluators) on public test data shows that the proposed model can perform text-simplifification at both lexical and syntactic levels, competitive to existing supervised methods. It also outperforms viable unsupervised baselines. Adding a few labeled pairs helps improve the performance further

上一篇:Towards Understanding Linear Word Analogies

下一篇:Variational Pretraining for Semi-supervised Text Classification

用户评价
全部评价

热门资源

  • Deep Cross-media ...

    Cross-media retrieval is a research hotspot in ...

  • Regularizing RNNs...

    Recently, caption generation with an encoder-de...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Joint Pose and Ex...

    Facial expression recognition (FER) is a challe...

  • Visual Reinforcem...

    For an autonomous agent to fulfill a wide range...