shReYas0363 commited on
Commit
68ca416
·
verified ·
1 Parent(s): e96df19

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -67
README.md CHANGED
@@ -1,67 +1,67 @@
1
- ---
2
- language:
3
- - en
4
- license: apache-2.0
5
- base_model: openai/whisper-tiny
6
- tags:
7
- - generated_from_trainer
8
- metrics:
9
- - wer
10
- model-index:
11
- - name: whispertiny-shreyas
12
- results: []
13
- ---
14
-
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
-
18
- # whispertiny-shreyas
19
-
20
- This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the AI4Bharat-svarah dataset.
21
- It achieves the following results on the evaluation set:
22
- - Loss: 0.5414
23
- - Wer: 22.8322
24
-
25
- ## Model description
26
-
27
- More information needed
28
-
29
- ## Intended uses & limitations
30
-
31
- More information needed
32
-
33
- ## Training and evaluation data
34
-
35
- More information needed
36
-
37
- ## Training procedure
38
-
39
- ### Training hyperparameters
40
-
41
- The following hyperparameters were used during training:
42
- - learning_rate: 1e-05
43
- - train_batch_size: 8
44
- - eval_batch_size: 8
45
- - seed: 42
46
- - gradient_accumulation_steps: 2
47
- - total_train_batch_size: 16
48
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
- - lr_scheduler_type: linear
50
- - lr_scheduler_warmup_steps: 500
51
- - training_steps: 2000
52
- - mixed_precision_training: Native AMP
53
-
54
- ### Training results
55
-
56
- | Training Loss | Epoch | Step | Validation Loss | Wer |
57
- |:-------------:|:------:|:----:|:---------------:|:-------:|
58
- | 0.2412 | 2.6702 | 1000 | 0.5319 | 22.8914 |
59
- | 0.1071 | 5.3405 | 2000 | 0.5414 | 22.8322 |
60
-
61
-
62
- ### Framework versions
63
-
64
- - Transformers 4.43.3
65
- - Pytorch 2.4.1
66
- - Datasets 2.14.7
67
- - Tokenizers 0.19.1
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ base_model: openai/whisper-tiny
6
+ tags:
7
+ - generated_from_trainer
8
+ metrics:
9
+ - wer
10
+ model-index:
11
+ - name: whispertiny-shreyas
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # whispertiny-shreyas
19
+
20
+ This model is a fine-tuned version of [Whisper Tiny](https://huggingface.co/openai/whisper-tiny) on the [AI4Bharat-svarah](https://github.com/AI4Bharat/Svarah) dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 0.5414
23
+ - Wer: 22.8322
24
+
25
+ ## Model description
26
+
27
+ Whisper is a Transformer based encoder-decoder model, also referred to as a sequence-to-sequence model. It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
28
+
29
+ The models were trained on either English-only data or multilingual data. The English-only models were trained on the task of speech recognition. The multilingual models were trained on both speech recognition and speech translation. For speech recognition, the model predicts transcriptions in the same language as the audio. For speech translation, the model predicts transcriptions to a different language to the audio.
30
+
31
+ Whisper checkpoints come in five configurations of varying model sizes. The smallest four are trained on either English-only or multilingual data. The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints are available on the Hugging Face Hub. The checkpoints are summarised in the following table with links to the models on the Hub:
32
+
33
+
34
+ ## Training procedure
35
+
36
+ Refer to [Sanchit's blog](https://huggingface.co/blog/fine-tune-whisper) and make changes according to the dependencies you have.
37
+
38
+
39
+ ### Training hyperparameters
40
+
41
+ The following hyperparameters were used during training:
42
+ - learning_rate: 1e-05
43
+ - train_batch_size: 8
44
+ - eval_batch_size: 8
45
+ - seed: 42
46
+ - gradient_accumulation_steps: 2
47
+ - total_train_batch_size: 16
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - lr_scheduler_warmup_steps: 500
51
+ - training_steps: 2000
52
+ - mixed_precision_training: Native AMP
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
57
+ |:-------------:|:------:|:----:|:---------------:|:-------:|
58
+ | 0.2412 | 2.6702 | 1000 | 0.5319 | 22.8914 |
59
+ | 0.1071 | 5.3405 | 2000 | 0.5414 | 22.8322 |
60
+
61
+
62
+ ### Framework versions
63
+
64
+ - Transformers 4.43.3
65
+ - Pytorch 2.4.1
66
+ - Datasets 2.14.7
67
+ - Tokenizers 0.19.1