KevinKibe commited on
Commit
73d0fb9
·
verified ·
1 Parent(s): 0fc6ed5

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- library_name: peft
3
  base_model: openai/whisper-small
 
4
  ---
5
 
6
  # Model Card for Model ID
 
1
  ---
 
2
  base_model: openai/whisper-small
3
+ library_name: peft
4
  ---
5
 
6
  # Model Card for Model ID
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:efef73abb8588e7e4a20c5916ae2fb82670cb01fabc7209d4b6c1fc7e9aa95fb
3
  size 14176064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a60f432e42c6f776d59999f121501a1583201a679d8c3db8793d4584a0cb887
3
  size 14176064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1163d7a9c1d5fdd03b75abf2f0a248b1ec482d3e105056719a0bdb2335d84b01
3
- size 4832890
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7dfdc64a80db410a3321f17f5a61c16c2e5343762d60a2ecf33b6a24e20bfbc
3
+ size 4832762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c79694f407c8b34fec70f8f58448c18e86ec8b13274047d470168fd99117ef22
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ba6c2f0d50b09ba332ede5e78d26a2b049d243564f39ab7379b4e173e564faa
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f17f17271f3b435ef6d1683459b3f439e4795b35f2028e8d831b629678dc9f52
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdb44de72ad26b6e9c5495c6369934563753532af0c4079339e42ec884246781
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,115 +1,35 @@
1
  {
2
- "best_metric": 110.00000000000001,
3
  "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-100",
4
- "epoch": 149.00666666666666,
5
- "eval_steps": 50,
6
- "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 24.006666666666668,
13
- "grad_norm": 1.3288557529449463,
14
- "learning_rate": 0.00026892857142857144,
15
- "loss": 3.2908,
16
- "step": 50
17
- },
18
- {
19
- "epoch": 24.006666666666668,
20
- "eval_loss": 2.3442695140838623,
21
- "eval_runtime": 65.8769,
22
- "eval_samples_per_second": 0.091,
23
- "eval_steps_per_second": 0.015,
24
- "eval_wer": 114.99999999999999,
25
- "step": 50
26
- },
27
- {
28
- "epoch": 49.00666666666667,
29
- "grad_norm": 0.20936653017997742,
30
- "learning_rate": 0.00021535714285714284,
31
- "loss": 0.8065,
32
  "step": 100
33
  },
34
  {
35
- "epoch": 49.00666666666667,
36
- "eval_loss": 2.699324607849121,
37
- "eval_runtime": 35.6683,
38
- "eval_samples_per_second": 0.168,
39
- "eval_steps_per_second": 0.028,
40
- "eval_wer": 110.00000000000001,
41
- "step": 100
42
- },
43
- {
44
- "epoch": 74.00666666666666,
45
- "grad_norm": 0.09886154532432556,
46
- "learning_rate": 0.00016178571428571427,
47
- "loss": 0.2445,
48
- "step": 150
49
- },
50
- {
51
- "epoch": 74.00666666666666,
52
- "eval_loss": 2.929717540740967,
53
- "eval_runtime": 36.3546,
54
- "eval_samples_per_second": 0.165,
55
- "eval_steps_per_second": 0.028,
56
- "eval_wer": 117.5,
57
- "step": 150
58
- },
59
- {
60
- "epoch": 99.00666666666666,
61
- "grad_norm": 0.056343283504247665,
62
- "learning_rate": 0.0001082142857142857,
63
- "loss": 0.0364,
64
- "step": 200
65
- },
66
- {
67
- "epoch": 99.00666666666666,
68
- "eval_loss": 3.129232168197632,
69
- "eval_runtime": 32.3636,
70
- "eval_samples_per_second": 0.185,
71
- "eval_steps_per_second": 0.031,
72
- "eval_wer": 114.99999999999999,
73
- "step": 200
74
- },
75
- {
76
- "epoch": 124.00666666666666,
77
- "grad_norm": 0.03805988281965256,
78
- "learning_rate": 5.4642857142857135e-05,
79
- "loss": 0.0144,
80
- "step": 250
81
- },
82
- {
83
- "epoch": 124.00666666666666,
84
- "eval_loss": 3.305183172225952,
85
- "eval_runtime": 34.8832,
86
- "eval_samples_per_second": 0.172,
87
  "eval_steps_per_second": 0.029,
88
- "eval_wer": 112.5,
89
- "step": 250
90
- },
91
- {
92
- "epoch": 149.00666666666666,
93
- "grad_norm": 0.03385389223694801,
94
- "learning_rate": 1.0714285714285714e-06,
95
- "loss": 0.0108,
96
- "step": 300
97
- },
98
- {
99
- "epoch": 149.00666666666666,
100
- "eval_loss": 3.287505865097046,
101
- "eval_runtime": 36.3654,
102
- "eval_samples_per_second": 0.165,
103
- "eval_steps_per_second": 0.027,
104
- "eval_wer": 110.00000000000001,
105
- "step": 300
106
  }
107
  ],
108
- "logging_steps": 50,
109
  "max_steps": 300,
110
  "num_input_tokens_seen": 0,
111
  "num_train_epochs": 9223372036854775807,
112
- "save_steps": 50,
113
  "stateful_callbacks": {
114
  "TrainerControl": {
115
  "args": {
@@ -117,12 +37,12 @@
117
  "should_evaluate": false,
118
  "should_log": false,
119
  "should_save": true,
120
- "should_training_stop": true
121
  },
122
  "attributes": {}
123
  }
124
  },
125
- "total_flos": 2.42287220736e+18,
126
  "train_batch_size": 32,
127
  "trial_name": null,
128
  "trial_params": null
 
1
  {
2
+ "best_metric": 112.23615784643621,
3
  "best_model_checkpoint": "../openai/whisper-small-finetuned/checkpoint-100",
4
+ "epoch": 1.1233333333333333,
5
+ "eval_steps": 100,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 1.1233333333333333,
13
+ "grad_norm": 1.4843510389328003,
14
+ "learning_rate": 7.214285714285714e-05,
15
+ "loss": 3.4293,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  "step": 100
17
  },
18
  {
19
+ "epoch": 1.1233333333333333,
20
+ "eval_loss": 3.470705986022949,
21
+ "eval_runtime": 448.6356,
22
+ "eval_samples_per_second": 0.892,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "eval_steps_per_second": 0.029,
24
+ "eval_wer": 112.23615784643621,
25
+ "step": 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  }
27
  ],
28
+ "logging_steps": 100,
29
  "max_steps": 300,
30
  "num_input_tokens_seen": 0,
31
  "num_train_epochs": 9223372036854775807,
32
+ "save_steps": 100,
33
  "stateful_callbacks": {
34
  "TrainerControl": {
35
  "args": {
 
37
  "should_evaluate": false,
38
  "should_log": false,
39
  "should_save": true,
40
+ "should_training_stop": false
41
  },
42
  "attributes": {}
43
  }
44
  },
45
+ "total_flos": 9.3508183130112e+17,
46
  "train_batch_size": 32,
47
  "trial_name": null,
48
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b625402a7f527e494329513826be1d2936e9c92c62e05bb2287e724674a4c79b
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abde47e545058c9f8627b6b9bd06a10d6ad5246e69db81c99e2d15a0e6a58642
3
  size 5368