pszemraj's picture
Update README.md
69715fe verified
|
raw
history blame
1.31 kB
---
library_name: transformers
license: apache-2.0
tags:
- tokenizer
- claude3
- t5
---
# claude3 tokenizer: for T5
Vocabulary size: 65103
- relevant special tokens for T5 training added
- post processor updated following t5's tokenizer
usage:
```py
from transformers import AutoTokenizer
tk = AutoTokenizer.from_pretrained('BEE-spoke-data/claude-tokenizer-forT5')
inputs = tk("here are some words", return_tensors="pt")
```
## post processor
```json
"post_processor": {
"type": "TemplateProcessing",
"single": [
{
"Sequence": {
"id": "A",
"type_id": 0
}
},
{
"SpecialToken": {
"id": "</s>",
"type_id": 0
}
}
],
"pair": [
{
"Sequence": {
"id": "A",
"type_id": 0
}
},
{
"SpecialToken": {
"id": "</s>",
"type_id": 0
}
},
{
"Sequence": {
"id": "B",
"type_id": 0
}
},
{
"SpecialToken": {
"id": "</s>",
"type_id": 0
}
}
],
"special_tokens": {
"</s>": {
"id": "</s>",
"ids": [
65001
],
"tokens": [
"</s>"
]
}
}
},
```