Upload tokenizer
Browse files- special_tokens_map.json +8 -1
- tokenizer_config.json +1 -1
special_tokens_map.json
CHANGED
@@ -108,5 +108,12 @@
|
|
108 |
"<|notimestamps|>"
|
109 |
],
|
110 |
"bos_token": "<|endoftext|>",
|
111 |
-
"eos_token": "<|endoftext|>"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
112 |
}
|
|
|
108 |
"<|notimestamps|>"
|
109 |
],
|
110 |
"bos_token": "<|endoftext|>",
|
111 |
+
"eos_token": "<|endoftext|>",
|
112 |
+
"unk_token": {
|
113 |
+
"content": "<|endoftext|>",
|
114 |
+
"lstrip": false,
|
115 |
+
"normalized": true,
|
116 |
+
"rstrip": false,
|
117 |
+
"single_word": false
|
118 |
+
}
|
119 |
}
|
tokenizer_config.json
CHANGED
@@ -19,7 +19,7 @@
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 1024,
|
22 |
-
"name_or_path": "/
|
23 |
"pad_token": null,
|
24 |
"processor_class": "WhisperProcessor",
|
25 |
"special_tokens_map_file": null,
|
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 1024,
|
22 |
+
"name_or_path": "openai/whisper-tiny",
|
23 |
"pad_token": null,
|
24 |
"processor_class": "WhisperProcessor",
|
25 |
"special_tokens_map_file": null,
|