alicegoesdown
commited on
Commit
•
b2b96ad
1
Parent(s):
20c2790
Training in progress, step 66, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2436967616
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17a581fe4340510e27eafe021c2fe545f48e7ece8d44cb412fb36a4c61793ff2
|
3 |
size 2436967616
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 671466706
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:677ab0db2ba4be7ff2cd83bfd86350c42a0588ae8dd0ce8ba95f57b5909abeb7
|
3 |
size 671466706
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3eba25965dd82562680a1b10ab4b72c93429b7d66978c078631695befb33f6ab
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3590c95beccaed6e985c8f99327ccb7e1e0604a90f734f2477dfbcb408474e61
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 4.000638484954834,
|
3 |
"best_model_checkpoint": "miner_id_24/checkpoint-50",
|
4 |
-
"epoch": 0.
|
5 |
"eval_steps": 25,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -381,6 +381,118 @@
|
|
381 |
"eval_samples_per_second": 3.424,
|
382 |
"eval_steps_per_second": 3.424,
|
383 |
"step": 50
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
384 |
}
|
385 |
],
|
386 |
"logging_steps": 1,
|
@@ -404,12 +516,12 @@
|
|
404 |
"should_evaluate": false,
|
405 |
"should_log": false,
|
406 |
"should_save": true,
|
407 |
-
"should_training_stop":
|
408 |
},
|
409 |
"attributes": {}
|
410 |
}
|
411 |
},
|
412 |
-
"total_flos": 1.
|
413 |
"train_batch_size": 1,
|
414 |
"trial_name": null,
|
415 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 4.000638484954834,
|
3 |
"best_model_checkpoint": "miner_id_24/checkpoint-50",
|
4 |
+
"epoch": 0.02107532032091965,
|
5 |
"eval_steps": 25,
|
6 |
+
"global_step": 66,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
381 |
"eval_samples_per_second": 3.424,
|
382 |
"eval_steps_per_second": 3.424,
|
383 |
"step": 50
|
384 |
+
},
|
385 |
+
{
|
386 |
+
"epoch": 0.01628547479343791,
|
387 |
+
"grad_norm": 18.63812255859375,
|
388 |
+
"learning_rate": 2.165719935902685e-05,
|
389 |
+
"loss": 3.9371,
|
390 |
+
"step": 51
|
391 |
+
},
|
392 |
+
{
|
393 |
+
"epoch": 0.01660479782860336,
|
394 |
+
"grad_norm": 15.783567428588867,
|
395 |
+
"learning_rate": 2.0214529598676836e-05,
|
396 |
+
"loss": 3.4496,
|
397 |
+
"step": 52
|
398 |
+
},
|
399 |
+
{
|
400 |
+
"epoch": 0.01692412086376881,
|
401 |
+
"grad_norm": 16.15505599975586,
|
402 |
+
"learning_rate": 1.8855661083370986e-05,
|
403 |
+
"loss": 3.1263,
|
404 |
+
"step": 53
|
405 |
+
},
|
406 |
+
{
|
407 |
+
"epoch": 0.01724344389893426,
|
408 |
+
"grad_norm": 12.968737602233887,
|
409 |
+
"learning_rate": 1.758386744638546e-05,
|
410 |
+
"loss": 3.0368,
|
411 |
+
"step": 54
|
412 |
+
},
|
413 |
+
{
|
414 |
+
"epoch": 0.017562766934099708,
|
415 |
+
"grad_norm": 9.380385398864746,
|
416 |
+
"learning_rate": 1.6402212549987762e-05,
|
417 |
+
"loss": 2.9326,
|
418 |
+
"step": 55
|
419 |
+
},
|
420 |
+
{
|
421 |
+
"epoch": 0.017882089969265158,
|
422 |
+
"grad_norm": 7.317039489746094,
|
423 |
+
"learning_rate": 1.531354310432403e-05,
|
424 |
+
"loss": 2.5816,
|
425 |
+
"step": 56
|
426 |
+
},
|
427 |
+
{
|
428 |
+
"epoch": 0.01820141300443061,
|
429 |
+
"grad_norm": 4.853557109832764,
|
430 |
+
"learning_rate": 1.4320481809445051e-05,
|
431 |
+
"loss": 2.6577,
|
432 |
+
"step": 57
|
433 |
+
},
|
434 |
+
{
|
435 |
+
"epoch": 0.018520736039596055,
|
436 |
+
"grad_norm": 4.9439239501953125,
|
437 |
+
"learning_rate": 1.3425421036992098e-05,
|
438 |
+
"loss": 3.1326,
|
439 |
+
"step": 58
|
440 |
+
},
|
441 |
+
{
|
442 |
+
"epoch": 0.018840059074761505,
|
443 |
+
"grad_norm": 4.261263847351074,
|
444 |
+
"learning_rate": 1.2630517066764069e-05,
|
445 |
+
"loss": 2.8784,
|
446 |
+
"step": 59
|
447 |
+
},
|
448 |
+
{
|
449 |
+
"epoch": 0.019159382109926956,
|
450 |
+
"grad_norm": 3.9718527793884277,
|
451 |
+
"learning_rate": 1.1937684892050604e-05,
|
452 |
+
"loss": 3.032,
|
453 |
+
"step": 60
|
454 |
+
},
|
455 |
+
{
|
456 |
+
"epoch": 0.019478705145092406,
|
457 |
+
"grad_norm": 3.4298315048217773,
|
458 |
+
"learning_rate": 1.1348593606245522e-05,
|
459 |
+
"loss": 2.6723,
|
460 |
+
"step": 61
|
461 |
+
},
|
462 |
+
{
|
463 |
+
"epoch": 0.019798028180257853,
|
464 |
+
"grad_norm": 3.560950994491577,
|
465 |
+
"learning_rate": 1.0864662381854632e-05,
|
466 |
+
"loss": 2.7488,
|
467 |
+
"step": 62
|
468 |
+
},
|
469 |
+
{
|
470 |
+
"epoch": 0.020117351215423303,
|
471 |
+
"grad_norm": 3.4607114791870117,
|
472 |
+
"learning_rate": 1.0487057051584856e-05,
|
473 |
+
"loss": 2.5312,
|
474 |
+
"step": 63
|
475 |
+
},
|
476 |
+
{
|
477 |
+
"epoch": 0.020436674250588753,
|
478 |
+
"grad_norm": 3.4417977333068848,
|
479 |
+
"learning_rate": 1.0216687299751144e-05,
|
480 |
+
"loss": 3.2348,
|
481 |
+
"step": 64
|
482 |
+
},
|
483 |
+
{
|
484 |
+
"epoch": 0.0207559972857542,
|
485 |
+
"grad_norm": 3.5609071254730225,
|
486 |
+
"learning_rate": 1.0054204470767243e-05,
|
487 |
+
"loss": 3.0496,
|
488 |
+
"step": 65
|
489 |
+
},
|
490 |
+
{
|
491 |
+
"epoch": 0.02107532032091965,
|
492 |
+
"grad_norm": 5.164323806762695,
|
493 |
+
"learning_rate": 1e-05,
|
494 |
+
"loss": 2.4215,
|
495 |
+
"step": 66
|
496 |
}
|
497 |
],
|
498 |
"logging_steps": 1,
|
|
|
516 |
"should_evaluate": false,
|
517 |
"should_log": false,
|
518 |
"should_save": true,
|
519 |
+
"should_training_stop": true
|
520 |
},
|
521 |
"attributes": {}
|
522 |
}
|
523 |
},
|
524 |
+
"total_flos": 1.969468656718971e+17,
|
525 |
"train_batch_size": 1,
|
526 |
"trial_name": null,
|
527 |
"trial_params": null
|