Bert Pretrained Token Embeddings
BERT(BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding) yields pretrained token (=subword) embeddings. Let's extract and save them in the word2vec format so that they can be used for downstream tasks.
Requirements
pytorch_pretrained_bert
NumPy
tqdm
Extraction
Bert (Pretrained) Token Embeddings in word2vec format
Example
* Related tokens to ##go