tbkazakova commited on
Commit
5add4a8
1 Parent(s): 7e464d5

End of training

Browse files
Files changed (4) hide show
  1. README.md +46 -10
  2. model.safetensors +1 -1
  3. preprocessor_config.json +0 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,13 +1,28 @@
1
  ---
2
  license: mit
 
3
  tags:
4
  - generated_from_trainer
5
- base_model: facebook/w2v-bert-2.0
6
  datasets:
7
  - audiofolder
 
 
8
  model-index:
9
  - name: wav2vec-bert-2.0-even-pakendorf-0406-1347
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -17,13 +32,9 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [facebook/w2v-bert-2.0](https://huggingface.co/facebook/w2v-bert-2.0) on the audiofolder dataset.
19
  It achieves the following results on the evaluation set:
20
- - eval_loss: inf
21
- - eval_wer: 0.9991
22
- - eval_runtime: 59.9347
23
- - eval_samples_per_second: 10.011
24
- - eval_steps_per_second: 1.251
25
- - epoch: 1.3333
26
- - step: 200
27
 
28
  ## Model description
29
 
@@ -54,9 +65,34 @@ The following hyperparameters were used during training:
54
  - num_epochs: 10
55
  - mixed_precision_training: Native AMP
56
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ### Framework versions
58
 
59
- - Transformers 4.41.1
60
  - Pytorch 2.3.0+cu121
61
  - Datasets 2.19.2
62
  - Tokenizers 0.19.1
 
1
  ---
2
  license: mit
3
+ base_model: facebook/w2v-bert-2.0
4
  tags:
5
  - generated_from_trainer
 
6
  datasets:
7
  - audiofolder
8
+ metrics:
9
+ - wer
10
  model-index:
11
  - name: wav2vec-bert-2.0-even-pakendorf-0406-1347
12
+ results:
13
+ - task:
14
+ name: Automatic Speech Recognition
15
+ type: automatic-speech-recognition
16
+ dataset:
17
+ name: audiofolder
18
+ type: audiofolder
19
+ config: default
20
+ split: train
21
+ args: default
22
+ metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 0.5968606805108706
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [facebook/w2v-bert-2.0](https://huggingface.co/facebook/w2v-bert-2.0) on the audiofolder dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Cer: 0.2128
36
+ - Loss: inf
37
+ - Wer: 0.5969
 
 
 
 
38
 
39
  ## Model description
40
 
 
65
  - num_epochs: 10
66
  - mixed_precision_training: Native AMP
67
 
68
+ ### Training results
69
+
70
+ | Training Loss | Epoch | Step | Cer | Validation Loss | Wer |
71
+ |:-------------:|:------:|:----:|:------:|:---------------:|:------:|
72
+ | 4.5767 | 0.5051 | 200 | 0.4932 | inf | 0.9973 |
73
+ | 1.8775 | 1.0101 | 400 | 0.3211 | inf | 0.8494 |
74
+ | 1.6006 | 1.5152 | 600 | 0.3017 | inf | 0.8040 |
75
+ | 1.4476 | 2.0202 | 800 | 0.2896 | inf | 0.7534 |
76
+ | 1.2213 | 2.5253 | 1000 | 0.2610 | inf | 0.7080 |
77
+ | 1.1485 | 3.0303 | 1200 | 0.2684 | inf | 0.6800 |
78
+ | 0.9554 | 3.5354 | 1400 | 0.2459 | inf | 0.6732 |
79
+ | 0.9379 | 4.0404 | 1600 | 0.2275 | inf | 0.6251 |
80
+ | 0.7644 | 4.5455 | 1800 | 0.2235 | inf | 0.6224 |
81
+ | 0.7891 | 5.0505 | 2000 | 0.2180 | inf | 0.6053 |
82
+ | 0.633 | 5.5556 | 2200 | 0.2130 | inf | 0.5996 |
83
+ | 0.6197 | 6.0606 | 2400 | 0.2126 | inf | 0.6032 |
84
+ | 0.5212 | 6.5657 | 2600 | 0.2196 | inf | 0.6019 |
85
+ | 0.4881 | 7.0707 | 2800 | 0.2125 | inf | 0.5894 |
86
+ | 0.4 | 7.5758 | 3000 | 0.2066 | inf | 0.5852 |
87
+ | 0.4008 | 8.0808 | 3200 | 0.2076 | inf | 0.5790 |
88
+ | 0.3304 | 8.5859 | 3400 | 0.2096 | inf | 0.5884 |
89
+ | 0.3446 | 9.0909 | 3600 | 0.2124 | inf | 0.5983 |
90
+ | 0.3237 | 9.5960 | 3800 | 0.2128 | inf | 0.5969 |
91
+
92
+
93
  ### Framework versions
94
 
95
+ - Transformers 4.41.2
96
  - Pytorch 2.3.0+cu121
97
  - Datasets 2.19.2
98
  - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b6e171d0a182489b9a3f59ba4ccf25019858e91ba2bf8c13b22301fd6499951
3
  size 2422978560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0e34fb80de4c4abc6da89c07bc2acca5c4a6333e215d3cea86419c4c616314d
3
  size 2422978560
preprocessor_config.json CHANGED
@@ -4,7 +4,6 @@
4
  "num_mel_bins": 80,
5
  "padding_side": "right",
6
  "padding_value": 0.0,
7
- "processor_class": "Wav2Vec2BertProcessor",
8
  "return_attention_mask": true,
9
  "sampling_rate": 16000,
10
  "stride": 2
 
4
  "num_mel_bins": 80,
5
  "padding_side": "right",
6
  "padding_value": 0.0,
 
7
  "return_attention_mask": true,
8
  "sampling_rate": 16000,
9
  "stride": 2
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e59af1a404864488356784a05d23870d7d8e06df6304dd544a5295723444e17a
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeeec4596f00f5d69e3fa4eea62251f33294831b4cb0dd1808f8662ed75ed1e7
3
  size 5112