ben81828 commited on
Commit
a4485bf
1 Parent(s): e71645a

Training in progress, step 1450, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2ea934ab728b1244cad904f211efc8dbea34bf120d0aaee69bc85b174ce604c
3
  size 18516456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c22365f9f1922257bce2b624ba131f902a872d3af6a64b3fc1e949516f00f5ac
3
  size 18516456
last-checkpoint/global_step1450/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c8959c735a1f2fc25bf1d4928b76efa4bb530eb830280ca22ca1633aa52d376
3
+ size 27700976
last-checkpoint/global_step1450/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1f85a672550122c013bfd5d71900a7266dd62f855af7c974a333f7a40dea95c
3
+ size 27700976
last-checkpoint/global_step1450/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f6cefdf65440a82655cd36f4399caf641f34ba34512af3342227658a10dd026
3
+ size 27700976
last-checkpoint/global_step1450/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9dbf6e1beb3b51e1eae928a14e8e1b7f54e2387d23032e91b3e09b668bb4f34
3
+ size 27700976
last-checkpoint/global_step1450/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bc9cf1986a83b78bb824f6697b3219dc70f628d204f6308b3f5c0110c8ec35a
3
+ size 411571
last-checkpoint/global_step1450/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa0aa15ab897602cd8484da3aeaeb273f468cc1ae87ee3cad1a86b4b8ac57b22
3
+ size 411507
last-checkpoint/global_step1450/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:223d06577f0627ab6d548c0af2814bcdad88b0a8c0c2157d9f9f6b236f33679a
3
+ size 411507
last-checkpoint/global_step1450/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39561dbca5b73760a417b8083b6ccb8825ee6b9f832540b6a585f4a97eef0a5a
3
+ size 411507
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step1400
 
1
+ global_step1450
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe66a68e61de2221b30fd9749bc68b45a1474bb2cc95901bca9557ac87909355
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd4f3298d54e9509917373bcd359e11c92a8e0aa77b2cc0825602efd186ad77e
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cd4f3162e46c3bb0f1fc4d3c52c7c33e60f56764458e0c8a73c3810b0a25f8c
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fdab02a791039ff46df83a272c972ab0f366fcea91338790dc016dbbbf6de80
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:185cc99aaa81b1b49b3ddc74aa6f97aa3036330983a7b69d52bd191057f9a5d5
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2808762a5f80587f05c1e2e8de26b5aef1010331cea366649319d593dbb0e66
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e37403c30cb4309e54e5defdde1906486716fc859274035d44aaac5d48a97ba
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c052112cbe3f325fd4543a02558091b80c9c8abcd8db8822be9a8306eab6f2b7
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba18a87eab7efca964506a3a18bdf9452eae09db2b3c77bd3a82db3283b5abc0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22e337e693b8d4ecad5d49a11b7f0649adec1748625a9390ba4ddabc0e901287
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.6643603444099426,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1400",
4
- "epoch": 0.7210919392222508,
5
  "eval_steps": 50,
6
- "global_step": 1400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2499,11 +2499,100 @@
2499
  "eval_steps_per_second": 0.927,
2500
  "num_input_tokens_seen": 16374784,
2501
  "step": 1400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2502
  }
2503
  ],
2504
  "logging_steps": 5,
2505
  "max_steps": 3400,
2506
- "num_input_tokens_seen": 16374784,
2507
  "num_train_epochs": 2,
2508
  "save_steps": 50,
2509
  "stateful_callbacks": {
@@ -2518,7 +2607,7 @@
2518
  "attributes": {}
2519
  }
2520
  },
2521
- "total_flos": 919473251942400.0,
2522
  "train_batch_size": 1,
2523
  "trial_name": null,
2524
  "trial_params": null
 
1
  {
2
  "best_metric": 0.6643603444099426,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1400",
4
+ "epoch": 0.7468452227659027,
5
  "eval_steps": 50,
6
+ "global_step": 1450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2499
  "eval_steps_per_second": 0.927,
2500
  "num_input_tokens_seen": 16374784,
2501
  "step": 1400
2502
+ },
2503
+ {
2504
+ "epoch": 0.723667267576616,
2505
+ "grad_norm": 2.665622720042704,
2506
+ "learning_rate": 6.806208330935766e-05,
2507
+ "loss": 0.706,
2508
+ "num_input_tokens_seen": 16433288,
2509
+ "step": 1405
2510
+ },
2511
+ {
2512
+ "epoch": 0.7262425959309812,
2513
+ "grad_norm": 2.123869663010538,
2514
+ "learning_rate": 6.783513316362855e-05,
2515
+ "loss": 0.6714,
2516
+ "num_input_tokens_seen": 16491784,
2517
+ "step": 1410
2518
+ },
2519
+ {
2520
+ "epoch": 0.7288179242853464,
2521
+ "grad_norm": 1.584213945279146,
2522
+ "learning_rate": 6.760776121436962e-05,
2523
+ "loss": 0.693,
2524
+ "num_input_tokens_seen": 16550272,
2525
+ "step": 1415
2526
+ },
2527
+ {
2528
+ "epoch": 0.7313932526397116,
2529
+ "grad_norm": 2.2481839233017764,
2530
+ "learning_rate": 6.737997283896103e-05,
2531
+ "loss": 0.7005,
2532
+ "num_input_tokens_seen": 16608704,
2533
+ "step": 1420
2534
+ },
2535
+ {
2536
+ "epoch": 0.7339685809940767,
2537
+ "grad_norm": 2.4818230151927643,
2538
+ "learning_rate": 6.715177342463145e-05,
2539
+ "loss": 0.6573,
2540
+ "num_input_tokens_seen": 16667200,
2541
+ "step": 1425
2542
+ },
2543
+ {
2544
+ "epoch": 0.7365439093484419,
2545
+ "grad_norm": 2.5398594354263486,
2546
+ "learning_rate": 6.692316836833065e-05,
2547
+ "loss": 0.6751,
2548
+ "num_input_tokens_seen": 16725704,
2549
+ "step": 1430
2550
+ },
2551
+ {
2552
+ "epoch": 0.7391192377028071,
2553
+ "grad_norm": 2.7486055345229343,
2554
+ "learning_rate": 6.6694163076602e-05,
2555
+ "loss": 0.6173,
2556
+ "num_input_tokens_seen": 16784192,
2557
+ "step": 1435
2558
+ },
2559
+ {
2560
+ "epoch": 0.7416945660571723,
2561
+ "grad_norm": 5.356237563459472,
2562
+ "learning_rate": 6.646476296545434e-05,
2563
+ "loss": 0.728,
2564
+ "num_input_tokens_seen": 16842704,
2565
+ "step": 1440
2566
+ },
2567
+ {
2568
+ "epoch": 0.7442698944115375,
2569
+ "grad_norm": 2.088505948846248,
2570
+ "learning_rate": 6.623497346023418e-05,
2571
+ "loss": 0.743,
2572
+ "num_input_tokens_seen": 16901176,
2573
+ "step": 1445
2574
+ },
2575
+ {
2576
+ "epoch": 0.7468452227659027,
2577
+ "grad_norm": 2.2198436340262,
2578
+ "learning_rate": 6.60047999954972e-05,
2579
+ "loss": 0.6291,
2580
+ "num_input_tokens_seen": 16959632,
2581
+ "step": 1450
2582
+ },
2583
+ {
2584
+ "epoch": 0.7468452227659027,
2585
+ "eval_loss": 0.753077507019043,
2586
+ "eval_runtime": 16.0383,
2587
+ "eval_samples_per_second": 3.741,
2588
+ "eval_steps_per_second": 0.935,
2589
+ "num_input_tokens_seen": 16959632,
2590
+ "step": 1450
2591
  }
2592
  ],
2593
  "logging_steps": 5,
2594
  "max_steps": 3400,
2595
+ "num_input_tokens_seen": 16959632,
2596
  "num_train_epochs": 2,
2597
  "save_steps": 50,
2598
  "stateful_callbacks": {
 
2607
  "attributes": {}
2608
  }
2609
  },
2610
+ "total_flos": 952313688883200.0,
2611
  "train_batch_size": 1,
2612
  "trial_name": null,
2613
  "trial_params": null