File size: 694 Bytes
bff929a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
from transformers import BertJapaneseTokenizer
from transformers import BertConfig
from transformers import BertForPreTraining
tokenizer = BertJapaneseTokenizer.from_pretrained('cl-tohoku/bert-base-japanese-whole-word-masking')
config = BertConfig().from_pretrained('cl-tohoku/bert-base-japanese-whole-word-masking')
tokenizer.save_pretrained('models/1-6_layer-wise')
config.save_pretrained('models/1-6_layer-wise')
tokenizer.save_pretrained('models/tapt512_60K')
config.save_pretrained('models/tapt512_60K')
tokenizer.save_pretrained('models/dapt128-tapt512')
config.save_pretrained('models/dapt128-tapt512')
# model = BertForPreTraining(config).from_pretrained('models/dapt128-tapt512')
|