drewschaub commited on
Commit
a86b0ca
1 Parent(s): c346999

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -34,9 +34,9 @@ should probably proofread and complete it, then remove this comment. -->
34
  # whisper-large-v3-japanese-4k-steps
35
 
36
  This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the Common Voice 16.1 dataset. I followed a post by Sanchit Gandhi, https://huggingface.co/blog/fine-tune-whisper
37
- It took 24 hours using an A100 on Google Colab to complete 4000 steps using the Common Voice 16.1 dataset. Training loss dropped over epochs but WER increased suggesting overfitting. It achieves the following results on the evaluation set:
38
  - Loss: 0.4057
39
- - Wer: 1821.4909
40
 
41
  ## Model description
42
 
@@ -71,10 +71,10 @@ The following hyperparameters were used during training:
71
 
72
  | Training Loss | Epoch | Step | Validation Loss | Wer |
73
  |:-------------:|:-----:|:----:|:---------------:|:---------:|
74
- | 0.1374 | 1.02 | 1000 | 0.3618 | 1198.3182 |
75
- | 0.0508 | 2.04 | 2000 | 0.3658 | 1755.4657 |
76
- | 0.0206 | 3.05 | 3000 | 0.3904 | 2108.7484 |
77
- | 0.0066 | 4.07 | 4000 | 0.4057 | 1821.4909 |
78
 
79
 
80
  ### Framework versions
 
34
  # whisper-large-v3-japanese-4k-steps
35
 
36
  This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the Common Voice 16.1 dataset. I followed a post by Sanchit Gandhi, https://huggingface.co/blog/fine-tune-whisper
37
+ It took 24 hours using an A100 on Google Colab to complete 4000 steps using the Common Voice 16.1 dataset. Training loss dropped over epochs but validation loss increased, so textbook overfitting. Furthermore, WER increased. It achieves the following results on the evaluation set:
38
  - Loss: 0.4057
39
+ - Wer: 18.2149
40
 
41
  ## Model description
42
 
 
71
 
72
  | Training Loss | Epoch | Step | Validation Loss | Wer |
73
  |:-------------:|:-----:|:----:|:---------------:|:---------:|
74
+ | 0.1374 | 1.02 | 1000 | 0.3618 | 11.983182 |
75
+ | 0.0508 | 2.04 | 2000 | 0.3658 | 17.554657 |
76
+ | 0.0206 | 3.05 | 3000 | 0.3904 | 21.087484 |
77
+ | 0.0066 | 4.07 | 4000 | 0.4057 | 18.214909 |
78
 
79
 
80
  ### Framework versions