trained-2k10-v4-model-merged / special_tokens_map.json
oMarquess's picture
Upload tokenizer
0efd1bd
raw
history blame
313 Bytes
{
"additional_special_tokens": [
">>TITLE<<",
">>ABSTRACT<<",
">>INTRODUCTION<<",
">>SUMMARY<<",
">>COMMENT<<",
">>ANSWER<<",
">>QUESTION<<",
">>DOMAIN<<",
">>PREFIX<<",
">>SUFFIX<<",
">>MIDDLE<<"
],
"eos_token": "<|endoftext|>",
"pad_token": "<|endoftext|>"
}