资源算法distill-bert

distill-bert

2020-03-10 | |  103 |   0 |   0

Distilled BERT

This work aims Knowledge Distillation from Google BERT model to compact Convolutional Models. (Not done yet)

Requirements

Python > 3.6, fire, tqdm, tensorboardx, tensorflow (for loading checkpoint file)

Example Usage

Fine-tuning (MRPC) Classifier with Pre-trained Transformer

Download BERT-Base, Uncased and GLUE Benchmark Datasets before fine-tuning.

  • make sure that "total_steps" in train.json should be greater than n_epochs*(num_data/batch_size)

Modify several config json files before following commands for training and evaluating.

python finetune.py config/finetune/mrpc/train.json
python finetune.py config/finetune/mrpc/eval.json

Training Blend CNN from scratch

See Transformer to CNN. Modify several config json files before following commands for training and evaluating.

python classify.py config/blendcnn/mrpc/train.json
python classify.py config/blendcnn/mrpc/eval.json

Knowledge Distillation from finetuned Transformer to CNN

Modify several config json files before following commands for training and evaluating.

python distill.py config/distill/mrpc/train.json
python distill.py config/distill/mrpc/eval.json




上一篇:bert-stack-overflow

下一篇: bert-token-embeddings

用户评价
全部评价

热门资源

  • DuReader_QANet_BiDAF

    Machine Reading Comprehension on DuReader Usin...

  • ETD_cataloguing_a...

    ETD catalouging project using allennlp

  • allennlp_extras

    allennlp_extras Some utilities build on top of...

  • allennlp-dureader

    An Apache 2.0 NLP research library, built on Py...

  • honk-honk-motherf...

    honk-honk-motherfucker