Update README.md
Browse files
README.md
CHANGED
@@ -19,11 +19,20 @@ This is a pruned version of [openai/whisper-base](https://huggingface.co/openai/
|
|
19 |
Pruning was made without any fine-tuning. Method from [this post](https://medium.com/m/global-identity-2?redirectUrl=https%3A%2F%2Ftowardsdatascience.com%2Fhow-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) was used.
|
20 |
|
21 |
## Size
|
22 |
-
Only 10% tokens was left
|
|
|
23 |
Model size is 30% less then original whisper-base:
|
24 |
| | openai/whisper-base | waveletdeboshir/whisper-base-ru-pruned |
|
25 |
| :------ | :------ | :------ |
|
26 |
| n of parameters | 74 M | 48.5 M |
|
27 |
| n of parameters (with proj_out layer) | 99 M | 51 M |
|
28 |
| model file size | 290 Mb | 203 Mb |
|
29 |
-
| vocab_size | 51865 | 4705 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
Pruning was made without any fine-tuning. Method from [this post](https://medium.com/m/global-identity-2?redirectUrl=https%3A%2F%2Ftowardsdatascience.com%2Fhow-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) was used.
|
20 |
|
21 |
## Size
|
22 |
+
Only 10% tokens was left including special whisper tokens, added whisper tokens, 100 most popular tokens from tokenizer and 3000 most popular Russian tokens computed by tokenization of russian text corpus.
|
23 |
+
|
24 |
Model size is 30% less then original whisper-base:
|
25 |
| | openai/whisper-base | waveletdeboshir/whisper-base-ru-pruned |
|
26 |
| :------ | :------ | :------ |
|
27 |
| n of parameters | 74 M | 48.5 M |
|
28 |
| n of parameters (with proj_out layer) | 99 M | 51 M |
|
29 |
| model file size | 290 Mb | 203 Mb |
|
30 |
+
| vocab_size | 51865 | 4705 |
|
31 |
+
|
32 |
+
## Metrics
|
33 |
+
Metrics for this model are on the same level as for openai/whisper-base.
|
34 |
+
|
35 |
+
You can fine-tune this model on your data to achive better performance.
|
36 |
+
|
37 |
+
## Colab for pruning
|
38 |
+
TODO
|