ben81828 commited on
Commit
459f888
·
verified ·
1 Parent(s): 149868f

Training in progress, step 2050, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b27b450bf69f26bb0a80206beeb4dd904291117c93705869b5a8ff84aa1e3f84
3
  size 18516456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e58273020b4ec80c27bd67a1dcf93114d440f82a76094d3bc50faf05938ff6dc
3
  size 18516456
last-checkpoint/global_step2049/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4659cb8935558c3f29594b6336fdf139feda949ae604a14b0ca0daf1dd0e39e8
3
+ size 27700976
last-checkpoint/global_step2049/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afe5cda623b4b40a90b10a4ffc064db2f9eeafa3c79959bf407d352bf536bd4d
3
+ size 27700976
last-checkpoint/global_step2049/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd333e9899a1e7c19b8c120109c9af616a96c14990e85766838bac9c5c801b66
3
+ size 27700976
last-checkpoint/global_step2049/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a2a05a345b53894b99bcca51f2d7b0bf5ddfe872c78fee6af0890fb76c759a5
3
+ size 27700976
last-checkpoint/global_step2049/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9086ce0cdbb623c2ce84fad441a76e576e34a80aec2724cc0323f604ead25f05
3
+ size 411571
last-checkpoint/global_step2049/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6f9680b16c527b3532e5f34fcddccd5d261758fbf64c1765a5583d2877490d6
3
+ size 411507
last-checkpoint/global_step2049/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2bb017dbec9d0f2a9dbc3e3548d74d100e3603cb3606daa0edf8ddf36497c1b
3
+ size 411507
last-checkpoint/global_step2049/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55416eff78220c5c9e94910c7dad6f64419db9d9c1fbfadc4d6f66ca6ca5b918
3
+ size 411507
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step1999
 
1
+ global_step2049
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4a695de1db3382235d3f8ae213672491aa2fdc3ba3be96403a089077ad3c2bf
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d1ddd7e7b4dc44903837b0414e4659f8383cd8f16b41dd396d4eaf5b9829f79
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2025b170fa1d4693537c2d73f89a6495c58940d033678742a74810c0154a6a7
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeddfb77fe4d3b495c4e08307767e08df90e96ef241c3eb80d5f75adec393e80
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0865c4d7d921b23a22c91c2f3b2c6cca03dae0eb27c43dee575c9602605c94d6
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e6260196fabb00061b1f1c8de6288382570dc14d02d2aa308050ca858880a97
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f4ec19d9df4417359523e8cc4d27875614c1021ebcc6391b27632aa7897b7ea
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea023bc5b1def54e0c49389175c0fae812f5f764c502525ce775d993d5ab2c03
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f900ab4b5112c328cc663d738cb777b204b3c8bdcbb2ceb91ed099a09d207ee1
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdc2918e410c169f1230e99c777b80638166c2769acb78c23195e4c48ca5b249
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.6319106221199036,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1600",
4
- "epoch": 1.029873808910636,
5
  "eval_steps": 50,
6
- "global_step": 2000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3567,11 +3567,100 @@
3567
  "eval_steps_per_second": 0.932,
3568
  "num_input_tokens_seen": 23386232,
3569
  "step": 2000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3570
  }
3571
  ],
3572
  "logging_steps": 5,
3573
  "max_steps": 3400,
3574
- "num_input_tokens_seen": 23386232,
3575
  "num_train_epochs": 2,
3576
  "save_steps": 50,
3577
  "stateful_callbacks": {
@@ -3586,7 +3675,7 @@
3586
  "attributes": {}
3587
  }
3588
  },
3589
- "total_flos": 1313241973129216.0,
3590
  "train_batch_size": 1,
3591
  "trial_name": null,
3592
  "trial_params": null
 
1
  {
2
  "best_metric": 0.6319106221199036,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1600",
4
+ "epoch": 1.0556270924542879,
5
  "eval_steps": 50,
6
+ "global_step": 2050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3567
  "eval_steps_per_second": 0.932,
3568
  "num_input_tokens_seen": 23386232,
3569
  "step": 2000
3570
+ },
3571
+ {
3572
+ "epoch": 1.0324491372650013,
3573
+ "grad_norm": 5.8532055715340245,
3574
+ "learning_rate": 3.938253580098613e-05,
3575
+ "loss": 0.362,
3576
+ "num_input_tokens_seen": 23444712,
3577
+ "step": 2005
3578
+ },
3579
+ {
3580
+ "epoch": 1.0350244656193666,
3581
+ "grad_norm": 7.087739461357715,
3582
+ "learning_rate": 3.914505045103845e-05,
3583
+ "loss": 0.3903,
3584
+ "num_input_tokens_seen": 23503192,
3585
+ "step": 2010
3586
+ },
3587
+ {
3588
+ "epoch": 1.0375997939737316,
3589
+ "grad_norm": 6.061997147134047,
3590
+ "learning_rate": 3.8907821822248605e-05,
3591
+ "loss": 0.3341,
3592
+ "num_input_tokens_seen": 23561688,
3593
+ "step": 2015
3594
+ },
3595
+ {
3596
+ "epoch": 1.0401751223280968,
3597
+ "grad_norm": 6.783069419644998,
3598
+ "learning_rate": 3.867085552510864e-05,
3599
+ "loss": 0.4794,
3600
+ "num_input_tokens_seen": 23620160,
3601
+ "step": 2020
3602
+ },
3603
+ {
3604
+ "epoch": 1.042750450682462,
3605
+ "grad_norm": 4.11088291372727,
3606
+ "learning_rate": 3.843415716390644e-05,
3607
+ "loss": 0.4104,
3608
+ "num_input_tokens_seen": 23678624,
3609
+ "step": 2025
3610
+ },
3611
+ {
3612
+ "epoch": 1.045325779036827,
3613
+ "grad_norm": 5.727855298190317,
3614
+ "learning_rate": 3.819773233659314e-05,
3615
+ "loss": 0.3639,
3616
+ "num_input_tokens_seen": 23737064,
3617
+ "step": 2030
3618
+ },
3619
+ {
3620
+ "epoch": 1.0479011073911924,
3621
+ "grad_norm": 6.936114108935384,
3622
+ "learning_rate": 3.7961586634650767e-05,
3623
+ "loss": 0.4294,
3624
+ "num_input_tokens_seen": 23795568,
3625
+ "step": 2035
3626
+ },
3627
+ {
3628
+ "epoch": 1.0504764357455576,
3629
+ "grad_norm": 5.577801320854008,
3630
+ "learning_rate": 3.772572564296005e-05,
3631
+ "loss": 0.4713,
3632
+ "num_input_tokens_seen": 23854040,
3633
+ "step": 2040
3634
+ },
3635
+ {
3636
+ "epoch": 1.0530517640999228,
3637
+ "grad_norm": 7.466883391944433,
3638
+ "learning_rate": 3.749015493966817e-05,
3639
+ "loss": 0.3864,
3640
+ "num_input_tokens_seen": 23912520,
3641
+ "step": 2045
3642
+ },
3643
+ {
3644
+ "epoch": 1.0556270924542879,
3645
+ "grad_norm": 4.120909561971508,
3646
+ "learning_rate": 3.7254880096057073e-05,
3647
+ "loss": 0.3602,
3648
+ "num_input_tokens_seen": 23971048,
3649
+ "step": 2050
3650
+ },
3651
+ {
3652
+ "epoch": 1.0556270924542879,
3653
+ "eval_loss": 0.7190810441970825,
3654
+ "eval_runtime": 16.0858,
3655
+ "eval_samples_per_second": 3.73,
3656
+ "eval_steps_per_second": 0.932,
3657
+ "num_input_tokens_seen": 23971048,
3658
+ "step": 2050
3659
  }
3660
  ],
3661
  "logging_steps": 5,
3662
  "max_steps": 3400,
3663
+ "num_input_tokens_seen": 23971048,
3664
  "num_train_epochs": 2,
3665
  "save_steps": 50,
3666
  "stateful_callbacks": {
 
3675
  "attributes": {}
3676
  }
3677
  },
3678
+ "total_flos": 1346083234512896.0,
3679
  "train_batch_size": 1,
3680
  "trial_name": null,
3681
  "trial_params": null