ben81828 commited on
Commit
a2a441f
1 Parent(s): 57bdcb6

Training in progress, step 1350, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69b791d45f5e8c5ec48a79f58002a2b81c7d9dc83eccb3702983e59a8069f0f5
3
  size 18516456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10b824529fcf1b1cc25b9074d2e972d61eae3cfa9d99a178c08c460258fd9cfb
3
  size 18516456
last-checkpoint/global_step1350/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1ff7758d634fd2817a944d2d4933749ce7e6d0682f36eaba1b11c022468402d
3
+ size 27700976
last-checkpoint/global_step1350/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:526d81af1a35b367eb6676a5cf443c017fa62dc7df664d3cb8fabbd164eefff7
3
+ size 27700976
last-checkpoint/global_step1350/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ed9346eb435cc9bef44166cf49d154244148637180ab0a5ac560e2cddfcc98b
3
+ size 27700976
last-checkpoint/global_step1350/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3649e71adc03e4a984ad48d9f9cc4d6f5cbf17ebf99ca7e2f85da1cc5aa0aee0
3
+ size 27700976
last-checkpoint/global_step1350/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba820a9b8f03145e1eef8bd5c6101f62091bf3ef869529103a40cefdc42a2d54
3
+ size 411571
last-checkpoint/global_step1350/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e32230362605afd218805da11ef95236d0ddb2e306753c377e8fabf7e581afb8
3
+ size 411507
last-checkpoint/global_step1350/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8bedad9cec5c02d9c769524a8aaba6c7bc518663b06904b7110383d2933a2cc
3
+ size 411507
last-checkpoint/global_step1350/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77c79e6a630a4d21da27b81765db87aa09442e7d1a1cdfe71098b67e6d64a5ae
3
+ size 411507
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step1300
 
1
+ global_step1350
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c364a094b8b4b8d6b015687012206e88b2233dd7d6a4f6f395d7aef77752ea67
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9747fe881253e52a47314f48068ef9649032bec4cb284b1b4becbb8787f37faa
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2bd2e297a13ac051fc5f3cce9c34767e51a5cb4574835aa8bd1309d8cdc48053
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ac02a5554a5ef9e3473dcd2926626ae41f4777354859c7d2bf0a0c1188c0583
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:028047dd1753d92e11bf971ee14a5c981a9a3ea6631f228e38475027eb5ae430
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9405d230cc78dac3f3b2ab887674631c15f66fedab0042ab7bc1bd83b8575344
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:000eb4cb4096582f15856f380d5f2a9a00eaecdbb95f2289a7a81a0a624fdf72
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b15237547030ac62d49d70a5465b2e29515e6334f62416eb16c0c6d073f7c6bf
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:480bbb635a5188e39c207a9dac4c52a057993fae728ac539c81e273941d4879d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca5b091f7e9fe97f38f540a5a26edd08efe79a114c6bc8ff4d4c0d9acb5b20f2
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.7189856171607971,
3
- "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1300",
4
- "epoch": 0.6695853721349472,
5
  "eval_steps": 50,
6
- "global_step": 1300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2321,11 +2321,100 @@
2321
  "eval_steps_per_second": 0.936,
2322
  "num_input_tokens_seen": 15205064,
2323
  "step": 1300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2324
  }
2325
  ],
2326
  "logging_steps": 5,
2327
  "max_steps": 3400,
2328
- "num_input_tokens_seen": 15205064,
2329
  "num_train_epochs": 2,
2330
  "save_steps": 50,
2331
  "stateful_callbacks": {
@@ -2340,7 +2429,7 @@
2340
  "attributes": {}
2341
  }
2342
  },
2343
- "total_flos": 853788671344640.0,
2344
  "train_batch_size": 1,
2345
  "trial_name": null,
2346
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.712917685508728,
3
+ "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1350",
4
+ "epoch": 0.695338655678599,
5
  "eval_steps": 50,
6
+ "global_step": 1350,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2321
  "eval_steps_per_second": 0.936,
2322
  "num_input_tokens_seen": 15205064,
2323
  "step": 1300
2324
+ },
2325
+ {
2326
+ "epoch": 0.6721607004893124,
2327
+ "grad_norm": 2.6154468701895066,
2328
+ "learning_rate": 7.250431889878455e-05,
2329
+ "loss": 0.7524,
2330
+ "num_input_tokens_seen": 15263560,
2331
+ "step": 1305
2332
+ },
2333
+ {
2334
+ "epoch": 0.6747360288436776,
2335
+ "grad_norm": 1.9549500311782502,
2336
+ "learning_rate": 7.228691778882693e-05,
2337
+ "loss": 0.6748,
2338
+ "num_input_tokens_seen": 15322016,
2339
+ "step": 1310
2340
+ },
2341
+ {
2342
+ "epoch": 0.6773113571980427,
2343
+ "grad_norm": 2.991178206089954,
2344
+ "learning_rate": 7.20689895899765e-05,
2345
+ "loss": 0.7571,
2346
+ "num_input_tokens_seen": 15380504,
2347
+ "step": 1315
2348
+ },
2349
+ {
2350
+ "epoch": 0.6798866855524079,
2351
+ "grad_norm": 1.7022848080804835,
2352
+ "learning_rate": 7.185053945626733e-05,
2353
+ "loss": 0.6615,
2354
+ "num_input_tokens_seen": 15438944,
2355
+ "step": 1320
2356
+ },
2357
+ {
2358
+ "epoch": 0.6824620139067731,
2359
+ "grad_norm": 1.739259284519112,
2360
+ "learning_rate": 7.163157255407732e-05,
2361
+ "loss": 0.7421,
2362
+ "num_input_tokens_seen": 15497384,
2363
+ "step": 1325
2364
+ },
2365
+ {
2366
+ "epoch": 0.6850373422611383,
2367
+ "grad_norm": 1.9142982939434143,
2368
+ "learning_rate": 7.141209406200599e-05,
2369
+ "loss": 0.7886,
2370
+ "num_input_tokens_seen": 15555856,
2371
+ "step": 1330
2372
+ },
2373
+ {
2374
+ "epoch": 0.6876126706155035,
2375
+ "grad_norm": 1.7562659805497576,
2376
+ "learning_rate": 7.1192109170752e-05,
2377
+ "loss": 0.7484,
2378
+ "num_input_tokens_seen": 15614368,
2379
+ "step": 1335
2380
+ },
2381
+ {
2382
+ "epoch": 0.6901879989698687,
2383
+ "grad_norm": 1.7590122465257017,
2384
+ "learning_rate": 7.097162308299054e-05,
2385
+ "loss": 0.7086,
2386
+ "num_input_tokens_seen": 15672864,
2387
+ "step": 1340
2388
+ },
2389
+ {
2390
+ "epoch": 0.6927633273242338,
2391
+ "grad_norm": 2.1211445265818845,
2392
+ "learning_rate": 7.07506410132501e-05,
2393
+ "loss": 0.7494,
2394
+ "num_input_tokens_seen": 15731376,
2395
+ "step": 1345
2396
+ },
2397
+ {
2398
+ "epoch": 0.695338655678599,
2399
+ "grad_norm": 2.683073565523052,
2400
+ "learning_rate": 7.052916818778918e-05,
2401
+ "loss": 0.7278,
2402
+ "num_input_tokens_seen": 15789848,
2403
+ "step": 1350
2404
+ },
2405
+ {
2406
+ "epoch": 0.695338655678599,
2407
+ "eval_loss": 0.712917685508728,
2408
+ "eval_runtime": 16.0726,
2409
+ "eval_samples_per_second": 3.733,
2410
+ "eval_steps_per_second": 0.933,
2411
+ "num_input_tokens_seen": 15789848,
2412
+ "step": 1350
2413
  }
2414
  ],
2415
  "logging_steps": 5,
2416
  "max_steps": 3400,
2417
+ "num_input_tokens_seen": 15789848,
2418
  "num_train_epochs": 2,
2419
  "save_steps": 50,
2420
  "stateful_callbacks": {
 
2429
  "attributes": {}
2430
  }
2431
  },
2432
+ "total_flos": 886628555423744.0,
2433
  "train_batch_size": 1,
2434
  "trial_name": null,
2435
  "trial_params": null