资源论文STRUCT BERT: INCORPORATING LANGUAGE STRUC -TURES INTO PRE -TRAINING FOR DEEP LANGUAGE UN -DERSTANDING

STRUCT BERT: INCORPORATING LANGUAGE STRUC -TURES INTO PRE -TRAINING FOR DEEP LANGUAGE UN -DERSTANDING

2019-12-31 | |  67 |   50 |   0

Abstract

Recently, the pre-trained language model, BERT (and its robustly optimized version RoBERTa), has attracted a lot of attention in natural language understanding (NLU), and achieved state-of-the-art accuracy in various NLU tasks, such as sentiment classification, natural language inference, semantic textual similarity and question answering. Inspired by the linearization exploration work of Elman (Elman, 1990), we extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the word and sentence levels, respectively. As a result, the new model is adapted to different levels of language understanding required by downstream tasks. The StructBERT with structural pre-training gives surprisingly good empirical results on a variety of downstream tasks, including pushing the state-of-the-art on the GLUE benchmark to 89.0 (outperforming all published models at the time of model submission), the F1 score on SQuAD v1.1 question answering to 93.0, the accuracy on SNLI to 91.7.

上一篇:VQ -WAV 2VEC :S ELF -S UPERVISED LEARNING OFD ISCRETE SPEECH REPRESENTATIONS

下一篇:LAMOL: LA NGUAGE MO DELING FORL IFELONG LANGUAGE LEARNING

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...