ben81828 commited on
Commit
cf33904
·
verified ·
1 Parent(s): bbd8539

Training in progress, step 200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1db5746fd65250f2bf6817d2de9266703177b7d0c3decef80aefa9773e3ca7a
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:381d89dd7c1135d0334349b9f49a436d03516c244b2ccee7000cd0fdc6d88805
3
  size 29034840
last-checkpoint/global_step200/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f73787e5bfc6d17b10cffea4f53d9c90b9a745d99e305b6688a004652359d006
3
+ size 43429616
last-checkpoint/global_step200/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9e8ea00d4eb5c712165d93ec8855b66f058a6461bf0b071821ec85f133579d6
3
+ size 43429616
last-checkpoint/global_step200/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a35a8adb32924b5460da62aff07373df5d13887e2b01f2bb77544ef5250fc9b
3
+ size 43429616
last-checkpoint/global_step200/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff6cff643243e7c2ce43ae1a8c95695a24e43038a042e26a11a9d99cb3ee00f4
3
+ size 43429616
last-checkpoint/global_step200/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c8aed679b2d5c060599974f1ab45390fb83c42d37e4ed63003977054cda112
3
+ size 637299
last-checkpoint/global_step200/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7ceb1154d5065a4632b1731df2acb5e51080a1f0388344212eaec5524452314
3
+ size 637171
last-checkpoint/global_step200/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30465b90efa0eb73fc80dc870a336c0cc691a34eaf088821a381a9f71ee7d652
3
+ size 637171
last-checkpoint/global_step200/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab524daca08828de12f8863c12ac5a32a74f179700a4a2850b77f7debcf459b3
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step150
 
1
+ global_step200
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70cc56408014c410353d4dd58ae9b03f4be043f5f800324f66fd8e20e99b840e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c5e18f922d0af74d820247ae97bee506ab412554a58345ddf2558abc94ee3e3
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49d1438e98cc9c53a6852464635ce62e9788e61eb3646b73e33813f487c4b6ae
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a2dcca6d9741f46592359768ea2212b9321da6408d1fd7d3a80b017bf37f434
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4388add9cec90932f8ff0100d27a0574d98e1bad52ff89d44e31967d2b4fbfde
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69420ece2c255923c5cbb3c6c9c4a6b9cb38fb57e5d3033c8b7d436a1faf6f13
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a705d6dfaae4f2c1b4b2be6b25a6eb521ffae6fcba21cc1531e97b60037ed079
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f278b40a1e23b88a657c4e5d03afa8dbbbe14dfeb16f6b4beedaece6cdd0b9
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c587ef55825cdfdbac47c29c3bd8a2996263c16833a5bd6aaa2fb014bea1e9d1
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdde69e82f45c9368fe627c2082d010eb77a8ceb0b2354f60caa089407a428a6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.09665286540985107,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_direction_scale4/lora/sft/checkpoint-150",
4
- "epoch": 0.07725985063095545,
5
  "eval_steps": 50,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -274,11 +274,100 @@
274
  "eval_steps_per_second": 0.778,
275
  "num_input_tokens_seen": 1497600,
276
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
277
  }
278
  ],
279
  "logging_steps": 5,
280
  "max_steps": 3400,
281
- "num_input_tokens_seen": 1497600,
282
  "num_train_epochs": 2,
283
  "save_steps": 50,
284
  "stateful_callbacks": {
@@ -293,7 +382,7 @@
293
  "attributes": {}
294
  }
295
  },
296
- "total_flos": 98764767100928.0,
297
  "train_batch_size": 1,
298
  "trial_name": null,
299
  "trial_params": null
 
1
  {
2
  "best_metric": 0.09665286540985107,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_direction_scale4/lora/sft/checkpoint-150",
4
+ "epoch": 0.10301313417460727,
5
  "eval_steps": 50,
6
+ "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
274
  "eval_steps_per_second": 0.778,
275
  "num_input_tokens_seen": 1497600,
276
  "step": 150
277
+ },
278
+ {
279
+ "epoch": 0.07983517898532062,
280
+ "grad_norm": 6.096163706748617,
281
+ "learning_rate": 9.11764705882353e-05,
282
+ "loss": 0.06,
283
+ "num_input_tokens_seen": 1547520,
284
+ "step": 155
285
+ },
286
+ {
287
+ "epoch": 0.08241050733968582,
288
+ "grad_norm": 4.278069142242893,
289
+ "learning_rate": 9.411764705882353e-05,
290
+ "loss": 0.0265,
291
+ "num_input_tokens_seen": 1597440,
292
+ "step": 160
293
+ },
294
+ {
295
+ "epoch": 0.08498583569405099,
296
+ "grad_norm": 1.7183097652953412,
297
+ "learning_rate": 9.705882352941177e-05,
298
+ "loss": 0.1743,
299
+ "num_input_tokens_seen": 1647360,
300
+ "step": 165
301
+ },
302
+ {
303
+ "epoch": 0.08756116404841617,
304
+ "grad_norm": 4.95401899568707,
305
+ "learning_rate": 0.0001,
306
+ "loss": 0.051,
307
+ "num_input_tokens_seen": 1697280,
308
+ "step": 170
309
+ },
310
+ {
311
+ "epoch": 0.09013649240278135,
312
+ "grad_norm": 0.34558354886099124,
313
+ "learning_rate": 9.999940874631277e-05,
314
+ "loss": 0.0584,
315
+ "num_input_tokens_seen": 1747200,
316
+ "step": 175
317
+ },
318
+ {
319
+ "epoch": 0.09271182075714654,
320
+ "grad_norm": 3.6326401692458146,
321
+ "learning_rate": 9.999763499923432e-05,
322
+ "loss": 0.0704,
323
+ "num_input_tokens_seen": 1797120,
324
+ "step": 180
325
+ },
326
+ {
327
+ "epoch": 0.09528714911151172,
328
+ "grad_norm": 0.4999720825488852,
329
+ "learning_rate": 9.999467880071402e-05,
330
+ "loss": 0.0278,
331
+ "num_input_tokens_seen": 1847040,
332
+ "step": 185
333
+ },
334
+ {
335
+ "epoch": 0.0978624774658769,
336
+ "grad_norm": 0.5211964505880501,
337
+ "learning_rate": 9.999054022066641e-05,
338
+ "loss": 0.0862,
339
+ "num_input_tokens_seen": 1896960,
340
+ "step": 190
341
+ },
342
+ {
343
+ "epoch": 0.10043780582024209,
344
+ "grad_norm": 0.8767038751832389,
345
+ "learning_rate": 9.998521935696953e-05,
346
+ "loss": 0.0565,
347
+ "num_input_tokens_seen": 1946880,
348
+ "step": 195
349
+ },
350
+ {
351
+ "epoch": 0.10301313417460727,
352
+ "grad_norm": 0.1848827361202722,
353
+ "learning_rate": 9.997871633546257e-05,
354
+ "loss": 0.0459,
355
+ "num_input_tokens_seen": 1996800,
356
+ "step": 200
357
+ },
358
+ {
359
+ "epoch": 0.10301313417460727,
360
+ "eval_loss": 0.09957947582006454,
361
+ "eval_runtime": 19.2651,
362
+ "eval_samples_per_second": 3.114,
363
+ "eval_steps_per_second": 0.779,
364
+ "num_input_tokens_seen": 1996800,
365
+ "step": 200
366
  }
367
  ],
368
  "logging_steps": 5,
369
  "max_steps": 3400,
370
+ "num_input_tokens_seen": 1996800,
371
  "num_train_epochs": 2,
372
  "save_steps": 50,
373
  "stateful_callbacks": {
 
382
  "attributes": {}
383
  }
384
  },
385
+ "total_flos": 131704746672128.0,
386
  "train_batch_size": 1,
387
  "trial_name": null,
388
  "trial_params": null