dada22231 commited on
Commit
fd6769e
1 Parent(s): a50dbfa

Training in progress, step 41, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74e800c35b11bfe1a08f96f05b791883d39e163009abc37ebeb8e92ff69b34f0
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61cede1e185ec7dbf624104fb40eb9fc0bc03a895e60dbede0c8424afa307e97
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:450b130cb7a2541a409ea797a151003a331b93d7d141da66fa847940f03534d6
3
  size 671466706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:119de6ebbe444568eb36723df3a483c1d702c8540c9fd99672a4810e73669d70
3
  size 671466706
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:410e2e49b2a3df2460526fdb53ed937cecd126897972ff2f8737560e70bfe1cf
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcb0eaa78b52c694c4941dc594e523bb92074f050c8e5b75b886bbdbc029cf8c
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ba1bb77e54d811be9e2894d17f12d5cb8eb4d1271696e61997d71cdeeb70690
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d19f5db0b5ed7101391ecc181bbc7740a892ec2d257f4fb3a05a75fb2983ffc
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f02f2c1321c3583d30f05af9c0b644720d402ddd65eaa5cc01bebd3e70b05d1f
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4f4bdf438cca51a87dc2daf2aeb8c9a8957897ddaddf4f5c2677f7a21a1a4b0
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e70944ff7fd3a11ea61949dbae7aa37388d84cd2c47931ef61cc9c9ed2b0233
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b591e0c41c4e20d058a924efcc112547252452a6f6cf09b413b46484fa398053
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:935c3511559349a68e0c75c3b87066ca7d4af7863dcb94ee59a533b76c75ad0c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a58533f5e61baaf0d295b9a19b1c4d2a6e1e9d96b361593f4d95a0f3cb76170
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7650135159492493,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-25",
4
- "epoch": 1.8648018648018647,
5
  "eval_steps": 25,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -198,6 +198,118 @@
198
  "eval_samples_per_second": 11.419,
199
  "eval_steps_per_second": 2.969,
200
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  }
202
  ],
203
  "logging_steps": 1,
@@ -221,12 +333,12 @@
221
  "should_evaluate": false,
222
  "should_log": false,
223
  "should_save": true,
224
- "should_training_stop": false
225
  },
226
  "attributes": {}
227
  }
228
  },
229
- "total_flos": 2.829013890367488e+17,
230
  "train_batch_size": 1,
231
  "trial_name": null,
232
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7650135159492493,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-25",
4
+ "epoch": 3.058275058275058,
5
  "eval_steps": 25,
6
+ "global_step": 41,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
198
  "eval_samples_per_second": 11.419,
199
  "eval_steps_per_second": 2.969,
200
  "step": 25
201
+ },
202
+ {
203
+ "epoch": 1.9393939393939394,
204
+ "grad_norm": 13.958155632019043,
205
+ "learning_rate": 3.904278008308589e-05,
206
+ "loss": 23.0799,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 2.013986013986014,
211
+ "grad_norm": 13.080740928649902,
212
+ "learning_rate": 3.570883473686256e-05,
213
+ "loss": 22.7616,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 2.0885780885780885,
218
+ "grad_norm": 9.918713569641113,
219
+ "learning_rate": 3.250000000000001e-05,
220
+ "loss": 22.2095,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 2.163170163170163,
225
+ "grad_norm": 12.069844245910645,
226
+ "learning_rate": 2.9437086397097995e-05,
227
+ "loss": 21.0234,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 2.237762237762238,
232
+ "grad_norm": 14.520157814025879,
233
+ "learning_rate": 2.6539958098208027e-05,
234
+ "loss": 21.4936,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 2.312354312354312,
239
+ "grad_norm": 9.91746711730957,
240
+ "learning_rate": 2.3827404092068032e-05,
241
+ "loss": 20.1627,
242
+ "step": 31
243
+ },
244
+ {
245
+ "epoch": 2.386946386946387,
246
+ "grad_norm": 11.826277732849121,
247
+ "learning_rate": 2.1317016332300447e-05,
248
+ "loss": 21.355,
249
+ "step": 32
250
+ },
251
+ {
252
+ "epoch": 2.4615384615384617,
253
+ "grad_norm": 13.523852348327637,
254
+ "learning_rate": 1.902507564684246e-05,
255
+ "loss": 20.8342,
256
+ "step": 33
257
+ },
258
+ {
259
+ "epoch": 2.5361305361305364,
260
+ "grad_norm": 11.003284454345703,
261
+ "learning_rate": 1.6966446150529244e-05,
262
+ "loss": 20.2875,
263
+ "step": 34
264
+ },
265
+ {
266
+ "epoch": 2.6107226107226107,
267
+ "grad_norm": 11.648452758789062,
268
+ "learning_rate": 1.515447884560556e-05,
269
+ "loss": 21.3661,
270
+ "step": 35
271
+ },
272
+ {
273
+ "epoch": 2.6853146853146854,
274
+ "grad_norm": 13.226274490356445,
275
+ "learning_rate": 1.3600925035352913e-05,
276
+ "loss": 20.4063,
277
+ "step": 36
278
+ },
279
+ {
280
+ "epoch": 2.7599067599067597,
281
+ "grad_norm": 12.421295166015625,
282
+ "learning_rate": 1.2315860112378455e-05,
283
+ "loss": 20.5653,
284
+ "step": 37
285
+ },
286
+ {
287
+ "epoch": 2.8344988344988344,
288
+ "grad_norm": 11.139906883239746,
289
+ "learning_rate": 1.130761821582766e-05,
290
+ "loss": 21.522,
291
+ "step": 38
292
+ },
293
+ {
294
+ "epoch": 2.909090909090909,
295
+ "grad_norm": 12.317776679992676,
296
+ "learning_rate": 1.0582738181293923e-05,
297
+ "loss": 21.0642,
298
+ "step": 39
299
+ },
300
+ {
301
+ "epoch": 2.983682983682984,
302
+ "grad_norm": 13.74180793762207,
303
+ "learning_rate": 1.0145921133960554e-05,
304
+ "loss": 20.5685,
305
+ "step": 40
306
+ },
307
+ {
308
+ "epoch": 3.058275058275058,
309
+ "grad_norm": 10.381887435913086,
310
+ "learning_rate": 1e-05,
311
+ "loss": 21.0001,
312
+ "step": 41
313
  }
314
  ],
315
  "logging_steps": 1,
 
333
  "should_evaluate": false,
334
  "should_log": false,
335
  "should_save": true,
336
+ "should_training_stop": true
337
  },
338
  "attributes": {}
339
  }
340
  },
341
+ "total_flos": 4.63958278020268e+17,
342
  "train_batch_size": 1,
343
  "trial_name": null,
344
  "trial_params": null