Arnold commited on
Commit
8c91de8
·
1 Parent(s): 5183b89

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -14
README.md CHANGED
@@ -15,9 +15,6 @@ should probably proofread and complete it, then remove this comment. -->
15
  # wav2vec2-large-xlsr-hausa2-demo-colab
16
 
17
  This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the common_voice dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.6617
20
- - Wer: 0.5962
21
 
22
  ## Model description
23
 
@@ -36,27 +33,20 @@ More information needed
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
- - learning_rate: 9.6e-05
40
  - train_batch_size: 16
41
  - eval_batch_size: 8
42
  - seed: 13
43
- - gradient_accumulation_steps: 2
44
- - total_train_batch_size: 32
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 500
48
- - num_epochs: 50
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
53
- | Training Loss | Epoch | Step | Validation Loss | Wer |
54
- |:-------------:|:-----:|:----:|:---------------:|:------:|
55
- | 6.9776 | 10.0 | 400 | 2.7735 | 1.0 |
56
- | 2.2732 | 20.0 | 800 | 0.7769 | 0.8212 |
57
- | 0.4471 | 30.0 | 1200 | 0.6199 | 0.6210 |
58
- | 0.2681 | 40.0 | 1600 | 0.6512 | 0.6022 |
59
- | 0.2191 | 50.0 | 2000 | 0.6617 | 0.5962 |
60
 
61
 
62
  ### Framework versions
 
15
  # wav2vec2-large-xlsr-hausa2-demo-colab
16
 
17
  This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the common_voice dataset.
 
 
 
18
 
19
  ## Model description
20
 
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
+ - learning_rate: 9.789e-05
37
  - train_batch_size: 16
38
  - eval_batch_size: 8
39
  - seed: 13
40
+ - gradient_accumulation_steps: 3
41
+ - total_train_batch_size: 48
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 500
45
+ - num_epochs: 40
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
 
 
 
 
 
 
 
50
 
51
 
52
  ### Framework versions