calvpang commited on
Commit
21b50d3
1 Parent(s): 1f84355

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -22,7 +22,7 @@ model-index:
22
  metrics:
23
  - name: Wer
24
  type: wer
25
- value: 31.83067619571193
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,9 +32,9 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the PolyAI/minds14 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.6024
36
- - Wer Ortho: 32.7775
37
- - Wer: 31.8307
38
 
39
  ## Model description
40
 
@@ -54,11 +54,11 @@ More information needed
54
 
55
  The following hyperparameters were used during training:
56
  - learning_rate: 1e-05
57
- - train_batch_size: 2
58
- - eval_batch_size: 2
59
  - seed: 42
60
- - gradient_accumulation_steps: 8
61
- - total_train_batch_size: 16
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: constant_with_warmup
64
  - lr_scheduler_warmup_steps: 50
@@ -66,14 +66,14 @@ The following hyperparameters were used during training:
66
 
67
  ### Training results
68
 
69
- | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
70
- |:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:|
71
- | 0.0007 | 17.86 | 500 | 0.6024 | 32.7775 | 31.8307 |
72
 
73
 
74
  ### Framework versions
75
 
76
  - Transformers 4.31.0
77
- - Pytorch 2.1.0.dev20230808
78
- - Datasets 2.14.3
79
  - Tokenizers 0.13.3
 
22
  metrics:
23
  - name: Wer
24
  type: wer
25
+ value: 0.3325867861142217
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the PolyAI/minds14 dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 0.7346
36
+ - Wer Ortho: 0.3489
37
+ - Wer: 0.3326
38
 
39
  ## Model description
40
 
 
54
 
55
  The following hyperparameters were used during training:
56
  - learning_rate: 1e-05
57
+ - train_batch_size: 16
58
+ - eval_batch_size: 16
59
  - seed: 42
60
+ - gradient_accumulation_steps: 4
61
+ - total_train_batch_size: 64
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: constant_with_warmup
64
  - lr_scheduler_warmup_steps: 50
 
66
 
67
  ### Training results
68
 
69
+ | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
70
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|
71
+ | 0.0004 | 71.43 | 500 | 0.7346 | 0.3489 | 0.3326 |
72
 
73
 
74
  ### Framework versions
75
 
76
  - Transformers 4.31.0
77
+ - Pytorch 2.0.1
78
+ - Datasets 2.14.4
79
  - Tokenizers 0.13.3