waveletdeboshir commited on
Commit
b3bdc93
·
verified ·
1 Parent(s): b329385

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -2
README.md CHANGED
@@ -19,11 +19,20 @@ This is a pruned version of [openai/whisper-base](https://huggingface.co/openai/
19
  Pruning was made without any fine-tuning. Method from [this post](https://medium.com/m/global-identity-2?redirectUrl=https%3A%2F%2Ftowardsdatascience.com%2Fhow-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) was used.
20
 
21
  ## Size
22
- Only 10% tokens was left: special whisper tokens, added whisper tokens, 100 most popular tokens from tokenizer and 3000 most popular Russian tokens computed by tokenization of russian text corpus.
 
23
  Model size is 30% less then original whisper-base:
24
  | | openai/whisper-base | waveletdeboshir/whisper-base-ru-pruned |
25
  | :------ | :------ | :------ |
26
  | n of parameters | 74 M | 48.5 M |
27
  | n of parameters (with proj_out layer) | 99 M | 51 M |
28
  | model file size | 290 Mb | 203 Mb |
29
- | vocab_size | 51865 | 4705 |
 
 
 
 
 
 
 
 
 
19
  Pruning was made without any fine-tuning. Method from [this post](https://medium.com/m/global-identity-2?redirectUrl=https%3A%2F%2Ftowardsdatascience.com%2Fhow-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) was used.
20
 
21
  ## Size
22
+ Only 10% tokens was left including special whisper tokens, added whisper tokens, 100 most popular tokens from tokenizer and 3000 most popular Russian tokens computed by tokenization of russian text corpus.
23
+
24
  Model size is 30% less then original whisper-base:
25
  | | openai/whisper-base | waveletdeboshir/whisper-base-ru-pruned |
26
  | :------ | :------ | :------ |
27
  | n of parameters | 74 M | 48.5 M |
28
  | n of parameters (with proj_out layer) | 99 M | 51 M |
29
  | model file size | 290 Mb | 203 Mb |
30
+ | vocab_size | 51865 | 4705 |
31
+
32
+ ## Metrics
33
+ Metrics for this model are on the same level as for openai/whisper-base.
34
+
35
+ You can fine-tune this model on your data to achive better performance.
36
+
37
+ ## Colab for pruning
38
+ TODO