patrickvonplaten
commited on
Commit
•
ce4eb45
1
Parent(s):
7872b1a
update tokenizer
Browse files- added_tokens.json +1 -0
- feature_extractor_config.json +0 -13
- preprocessor_config.json +15 -6
- special_tokens_map.json +1 -1
- tokenizer_config.json +1 -1
- vocab.json +0 -0
added_tokens.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"<s>": 8065, "</s>": 8066}
|
feature_extractor_config.json
DELETED
@@ -1,13 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"do_normalize": true,
|
3 |
-
"feature_size": 80,
|
4 |
-
"sampling_rate": 16000,
|
5 |
-
"padding_value": 0.0,
|
6 |
-
"hop_length": 10,
|
7 |
-
"win_length": 25,
|
8 |
-
"win_function": "hamming_window",
|
9 |
-
"normalize_means": true,
|
10 |
-
"normalize_vars": true,
|
11 |
-
"return_attention_mask": false
|
12 |
-
}
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
preprocessor_config.json
CHANGED
@@ -1,12 +1,21 @@
|
|
1 |
{
|
|
|
2 |
"do_normalize": true,
|
|
|
3 |
"feature_size": 80,
|
4 |
-
"
|
5 |
-
"padding_value": 0.0,
|
6 |
"hop_length": 10,
|
7 |
-
"
|
8 |
-
"
|
9 |
"normalize_means": true,
|
10 |
-
"normalize_vars": true,
|
11 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
}
|
|
|
1 |
{
|
2 |
+
"K": 257,
|
3 |
"do_normalize": true,
|
4 |
+
"feature_extractor_type": "MCTCFeatureExtractor",
|
5 |
"feature_size": 80,
|
6 |
+
"frame_signal_scale": 32768.0,
|
|
|
7 |
"hop_length": 10,
|
8 |
+
"mel_floor": 1.0,
|
9 |
+
"n_fft": 512,
|
10 |
"normalize_means": true,
|
11 |
+
"normalize_vars": true,
|
12 |
+
"padding_side": "right",
|
13 |
+
"padding_value": 0.0,
|
14 |
+
"preemphasis_coeff": 0.97,
|
15 |
+
"return_attention_mask": false,
|
16 |
+
"sample_size": 400,
|
17 |
+
"sample_stride": 160,
|
18 |
+
"sampling_rate": 16000,
|
19 |
+
"win_function": "hamming_window",
|
20 |
+
"win_length": 25
|
21 |
}
|
special_tokens_map.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "return_attention_mask": false, "do_normalize": true}
|
|
|
1 |
+
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "do_lower_case": false, "word_delimiter_token": "|", "replace_word_delimiter_char": " ", "return_attention_mask": false, "do_normalize": true, "special_tokens_map_file": "./mctc-large/special_tokens_map.json", "name_or_path": "./mctc-large", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
|
vocab.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|