Home

Awesome

Bert Pretrained Token Embeddings

BERT(BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding) yields pretrained token (=subword) embeddings. Let's extract and save them in the word2vec format so that they can be used for downstream tasks.

Requirements

Extraction

Bert (Pretrained) Token Embeddings in word2vec format

Models# Vocab# DimNotes
bert-base-uncased30,522768
bert-large-uncased30,5221024
bert-base-cased28,996768
bert-large-cased28,9961024
bert-base-multilingual-cased119,547768Recommended
bert-base-multilingual-uncased30,522768Not recommended
bert-base-chinese21,128768

Example

<img src="go.png" width=1000>