dzanbek commited on
Commit
2e0db63
·
verified ·
1 Parent(s): 66e920c

Training in progress, step 36, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8fdab2749ab700cdcbd215765631960a327e512cd01b5f04b030e416e065972
3
  size 50624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcd7ffcb3e4a9f4bfc4ad8648cfbb675bd6afe0bd814d648853e775947350c0f
3
  size 50624
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ccc9927edd24a80b113f328d7e28b70cd0fb609083e99505be64422e8b327f6
3
  size 118090
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4928a4ea058d36d00d5fa931f38e737bf9e77813884c198e504dd55291d5efaf
3
  size 118090
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5937b2d23ad85f37009cf032d023af240d8787d5d33f30bae27a2bd090137a04
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a6d12bbdf31e66e64bbfd314e06ec60bdc4d273af3264013c74689308316443
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06c69db2db6de56f38ba12b474a491d20087e27dc2893a95d6ac7716476ca645
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aa0f75a1f3e346be25756b578158b09a68943f0b9f1cfe29f97939687f864ef
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.16120906801007556,
5
  "eval_steps": 4,
6
- "global_step": 32,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -303,6 +303,42 @@
303
  "eval_samples_per_second": 128.109,
304
  "eval_steps_per_second": 64.054,
305
  "step": 32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
306
  }
307
  ],
308
  "logging_steps": 1,
@@ -322,7 +358,7 @@
322
  "attributes": {}
323
  }
324
  },
325
- "total_flos": 6518401400832.0,
326
  "train_batch_size": 2,
327
  "trial_name": null,
328
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.181360201511335,
5
  "eval_steps": 4,
6
+ "global_step": 36,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
303
  "eval_samples_per_second": 128.109,
304
  "eval_steps_per_second": 64.054,
305
  "step": 32
306
+ },
307
+ {
308
+ "epoch": 0.16624685138539042,
309
+ "grad_norm": 0.1841493397951126,
310
+ "learning_rate": 7.66554636144095e-05,
311
+ "loss": 11.7572,
312
+ "step": 33
313
+ },
314
+ {
315
+ "epoch": 0.1712846347607053,
316
+ "grad_norm": 0.1587410420179367,
317
+ "learning_rate": 6.909830056250527e-05,
318
+ "loss": 11.7721,
319
+ "step": 34
320
+ },
321
+ {
322
+ "epoch": 0.17632241813602015,
323
+ "grad_norm": 0.1748318076133728,
324
+ "learning_rate": 6.173165676349103e-05,
325
+ "loss": 11.7697,
326
+ "step": 35
327
+ },
328
+ {
329
+ "epoch": 0.181360201511335,
330
+ "grad_norm": 0.15539947152137756,
331
+ "learning_rate": 5.4600950026045326e-05,
332
+ "loss": 11.7722,
333
+ "step": 36
334
+ },
335
+ {
336
+ "epoch": 0.181360201511335,
337
+ "eval_loss": 11.765742301940918,
338
+ "eval_runtime": 0.6575,
339
+ "eval_samples_per_second": 127.75,
340
+ "eval_steps_per_second": 63.875,
341
+ "step": 36
342
  }
343
  ],
344
  "logging_steps": 1,
 
358
  "attributes": {}
359
  }
360
  },
361
+ "total_flos": 7333201575936.0,
362
  "train_batch_size": 2,
363
  "trial_name": null,
364
  "trial_params": null