Kavya Manohar commited on
Commit
fef0c48
·
1 Parent(s): 21fe2d4

End of training

Browse files
Files changed (4) hide show
  1. README.md +18 -45
  2. generation_config.json +17 -69
  3. pytorch_model.bin +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,40 +1,23 @@
1
  ---
2
  language:
3
  - ml
4
- license: mit
 
5
  tags:
6
- - whisper-event
7
- - hf-asr-leaderboard
8
  - generated_from_trainer
9
  datasets:
10
- - mozilla-foundation/common_voice_11_0
11
- metrics:
12
- - wer
13
- base_model: openai/whisper-small
14
  model-index:
15
- - name: whisper_malayalam_largev2
16
- results:
17
- - task:
18
- type: automatic-speech-recognition
19
- name: Automatic Speech Recognition
20
- dataset:
21
- name: mozilla-foundation/common_voice_11_0
22
- type: mozilla-foundation/common_voice_11_0
23
- config: ml
24
- split: test
25
- metrics:
26
- - type: wer
27
- value: 84.36781609195401
28
- name: Wer
29
  ---
30
 
31
- # Whisper-fineTuning-malayalam
 
32
 
33
- This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset.
34
- It achieves the following results on the evaluation set:
35
 
36
- Loss: 0.5664
37
- Wer: 84.3678
38
 
39
  ## Model description
40
 
@@ -54,27 +37,17 @@ More information needed
54
 
55
  The following hyperparameters were used during training:
56
  - learning_rate: 1e-05
57
- - train_batch_size: 12
58
- - eval_batch_size: 8
59
  - seed: 42
60
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
61
- - lr_scheduler_type: linear
62
- - lr_scheduler_warmup_steps: 500
63
- - training_steps: 1000
64
- - mixed_precision_training: Native AMP
65
-
66
- ### Training results
67
-
68
- 27.78 1000
69
-
70
- | Training Loss | Epoch | Step | Validation Loss | Wer |
71
- |:-------------:|:-----:|:----:|:---------------:|:-------:|
72
- | 0.0003 | 27.78 | 1000 | 0.5664 | 84.3678 |
73
-
74
 
75
  ### Framework versions
76
 
77
- - Transformers 4.26.0.dev0
78
- - Pytorch 1.13.0+cu116
79
- - Datasets 2.8.0
80
- - Tokenizers 0.13.2
 
1
  ---
2
  language:
3
  - ml
4
+ license: apache-2.0
5
+ base_model: openai/whisper-small
6
  tags:
 
 
7
  - generated_from_trainer
8
  datasets:
9
+ - mozilla-foundation/common_voice_13_0
 
 
 
10
  model-index:
11
+ - name: Whisper Small Malayalam - Kavya Manohar
12
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
 
18
+ # Whisper Small Malayalam - Kavya Manohar
 
19
 
20
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 13 dataset.
 
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
+ - train_batch_size: 16
41
+ - eval_batch_size: 16
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: constant_with_warmup
45
+ - lr_scheduler_warmup_steps: 50
46
+ - training_steps: 500
 
 
 
 
 
 
 
 
 
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.33.3
51
+ - Pytorch 2.0.1+cu118
52
+ - Datasets 2.14.5
53
+ - Tokenizers 0.13.3
generation_config.json CHANGED
@@ -1,96 +1,44 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 10,
5
- 12
6
- ],
7
- [
8
- 13,
9
- 17
10
- ],
11
- [
12
- 16,
13
- 11
14
- ],
15
- [
16
- 16,
17
- 12
18
- ],
19
- [
20
- 16,
21
- 13
22
- ],
23
- [
24
- 17,
25
- 15
26
- ],
27
- [
28
- 17,
29
- 16
30
- ],
31
- [
32
- 18,
33
- 4
34
- ],
35
- [
36
- 18,
37
- 11
38
- ],
39
- [
40
- 18,
41
- 19
42
- ],
43
- [
44
- 19,
45
- 11
46
- ],
47
- [
48
- 21,
49
- 2
50
- ],
51
- [
52
- 21,
53
- 3
54
- ],
55
- [
56
- 22,
57
  3
58
  ],
59
  [
60
- 22,
61
  9
62
  ],
63
  [
64
- 22,
65
- 12
66
  ],
67
  [
68
- 23,
69
- 5
70
  ],
71
  [
72
- 23,
73
  7
74
  ],
75
  [
76
- 23,
77
- 13
78
  ],
79
  [
80
- 25,
81
- 5
82
  ],
83
  [
84
- 26,
85
- 1
86
  ],
87
  [
88
- 26,
89
- 12
90
  ],
91
  [
92
- 27,
93
- 15
94
  ]
95
  ],
96
  "begin_suppress_tokens": [
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 5,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  3
6
  ],
7
  [
8
+ 5,
9
  9
10
  ],
11
  [
12
+ 8,
13
+ 0
14
  ],
15
  [
16
+ 8,
17
+ 4
18
  ],
19
  [
20
+ 8,
21
  7
22
  ],
23
  [
24
+ 8,
25
+ 8
26
  ],
27
  [
28
+ 9,
29
+ 0
30
  ],
31
  [
32
+ 9,
33
+ 7
34
  ],
35
  [
36
+ 9,
37
+ 9
38
  ],
39
  [
40
+ 10,
41
+ 5
42
  ]
43
  ],
44
  "begin_suppress_tokens": [
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17c9574ff2b03a12db1679e0bed9c65ff6068b34d6f77e369cd76cabf043a515
3
- size 967102729
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55b01a43f8ad46754c18f9ec9fe2e066ef91975863b135699b07edbb0c0cff28
3
+ size 967096009
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45a90ce4dea18bb0faa363deb0cd2146948a0f17c5b9a0e284d1d317069e312f
3
  size 4219
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e76e435e8d0911d081c337a92be220eb31366515caedec24857a54154ab14c7
3
  size 4219