ben81828 commited on
Commit
3ac0017
·
verified ·
1 Parent(s): 260589e

Training in progress, step 450, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5682b2eafe76e843b3c252d9efdff6d8e222ea2c09fcb985ed44476c6c49d3c0
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e3aaf9e6c8d830995b2a237a6a326491ecc141ce2f9713c7cafa4ab230f17cb
3
  size 29034840
last-checkpoint/global_step450/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5c814ad588b1665f2b04d0687e9a2e36e4add22db1e7d4a157a85a1e5dcbc00
3
+ size 43429616
last-checkpoint/global_step450/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9d1bf853471ab45c77df069f67590378cc45e763121492de3a880e050e5387b
3
+ size 43429616
last-checkpoint/global_step450/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67c7c9115f51b9a82946db16ad5410c547718f1a13c8703a0865244932e60537
3
+ size 43429616
last-checkpoint/global_step450/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85a9db1b2a23e12f16a96377f5b5f03a144a90e62ecc5a6634e7ee8832c81cd1
3
+ size 43429616
last-checkpoint/global_step450/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2c43b6581640ec9140a8e671944ff69d96a0284c4069ddbe5a73547523d050e
3
+ size 637299
last-checkpoint/global_step450/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e7b078653231149a903bc00280c61fc902a6ebe6078af02817b5efd13e9ef2b
3
+ size 637171
last-checkpoint/global_step450/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3014bb4485c273ec85ee6adab304455dcd75b99ca668f9c10e04f1ebc443fec8
3
+ size 637171
last-checkpoint/global_step450/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f23e7a4eb8e027c3a1e9ed72ab3097610a2c4faa00790d4b003da7cdb06018e2
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step400
 
1
+ global_step450
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:543ef05f530d40ee20b8d626b07a69b86597aca643e48897571062f973efe84f
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dbc6521b0b64cb12d818506108fcf257a4089ca8a9b1e453776ed3e032e7176
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a23f732e43838ce0398d2636885ac16badbb9bcbc04d1406069ba3027bc5ae0
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b13e3da1b0679cab1bab94f893e385a9a224d3335b5a6f62602f33c2be88d03
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e10cce960e7068b051c05e35ed6160656be9091c63f13796ac2ed7e9c84e5a72
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a24f0e0f117b5a8236e0d12594c0c358f41ef00068d4460002e95ad1cc3cb1c
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6f6049e212b1df5cefc5d834afcd8cc052c73f1457449e9fe8a38d514f54078
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e46e4eab6c4a25d84ad36ddf1357401788adeeb6388c03cefa35a63b52ee7610
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4fa1113b4807f273646ff4a5c640f26a2f072e0ac3eba31e530c029f936b8156
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:312387a763149a43bd97e3a38958946d432aa6037fb7604fcd6a6729912e1b32
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7038857936859131,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_direction_then_DetectAndClassify_scale6/lora/sft/checkpoint-400",
4
- "epoch": 0.11816838995568685,
5
  "eval_steps": 50,
6
- "global_step": 400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -719,11 +719,100 @@
719
  "eval_steps_per_second": 0.787,
720
  "num_input_tokens_seen": 4142592,
721
  "step": 400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
722
  }
723
  ],
724
  "logging_steps": 5,
725
  "max_steps": 6770,
726
- "num_input_tokens_seen": 4142592,
727
  "num_train_epochs": 2,
728
  "save_steps": 50,
729
  "stateful_callbacks": {
@@ -738,7 +827,7 @@
738
  "attributes": {}
739
  }
740
  },
741
- "total_flos": 273245789749248.0,
742
  "train_batch_size": 1,
743
  "trial_name": null,
744
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7038857936859131,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_direction_then_DetectAndClassify_scale6/lora/sft/checkpoint-400",
4
+ "epoch": 0.1329394387001477,
5
  "eval_steps": 50,
6
+ "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
719
  "eval_steps_per_second": 0.787,
720
  "num_input_tokens_seen": 4142592,
721
  "step": 400
722
+ },
723
+ {
724
+ "epoch": 0.11964549483013294,
725
+ "grad_norm": 4.400494365327609,
726
+ "learning_rate": 9.997401438591772e-05,
727
+ "loss": 0.6209,
728
+ "num_input_tokens_seen": 4194920,
729
+ "step": 405
730
+ },
731
+ {
732
+ "epoch": 0.12112259970457903,
733
+ "grad_norm": 4.518582133930376,
734
+ "learning_rate": 9.996992844441495e-05,
735
+ "loss": 0.6576,
736
+ "num_input_tokens_seen": 4247048,
737
+ "step": 410
738
+ },
739
+ {
740
+ "epoch": 0.12259970457902511,
741
+ "grad_norm": 2.6773114089558043,
742
+ "learning_rate": 9.996554438282022e-05,
743
+ "loss": 0.6851,
744
+ "num_input_tokens_seen": 4299728,
745
+ "step": 415
746
+ },
747
+ {
748
+ "epoch": 0.1240768094534712,
749
+ "grad_norm": 10.583849604294256,
750
+ "learning_rate": 9.996086222728879e-05,
751
+ "loss": 0.6288,
752
+ "num_input_tokens_seen": 4351088,
753
+ "step": 420
754
+ },
755
+ {
756
+ "epoch": 0.1255539143279173,
757
+ "grad_norm": 5.18430085456359,
758
+ "learning_rate": 9.995588200575439e-05,
759
+ "loss": 0.667,
760
+ "num_input_tokens_seen": 4403016,
761
+ "step": 425
762
+ },
763
+ {
764
+ "epoch": 0.12703101920236337,
765
+ "grad_norm": 3.624202284960618,
766
+ "learning_rate": 9.995060374792892e-05,
767
+ "loss": 0.6747,
768
+ "num_input_tokens_seen": 4453880,
769
+ "step": 430
770
+ },
771
+ {
772
+ "epoch": 0.12850812407680945,
773
+ "grad_norm": 8.193068077958594,
774
+ "learning_rate": 9.994502748530244e-05,
775
+ "loss": 0.6594,
776
+ "num_input_tokens_seen": 4505616,
777
+ "step": 435
778
+ },
779
+ {
780
+ "epoch": 0.12998522895125553,
781
+ "grad_norm": 6.0681139183306145,
782
+ "learning_rate": 9.993915325114288e-05,
783
+ "loss": 0.6727,
784
+ "num_input_tokens_seen": 4558384,
785
+ "step": 440
786
+ },
787
+ {
788
+ "epoch": 0.13146233382570163,
789
+ "grad_norm": 3.998790148445953,
790
+ "learning_rate": 9.993298108049582e-05,
791
+ "loss": 0.6526,
792
+ "num_input_tokens_seen": 4611184,
793
+ "step": 445
794
+ },
795
+ {
796
+ "epoch": 0.1329394387001477,
797
+ "grad_norm": 3.1838689643423392,
798
+ "learning_rate": 9.992651101018445e-05,
799
+ "loss": 0.5661,
800
+ "num_input_tokens_seen": 4663320,
801
+ "step": 450
802
+ },
803
+ {
804
+ "epoch": 0.1329394387001477,
805
+ "eval_loss": 0.7132604718208313,
806
+ "eval_runtime": 18.996,
807
+ "eval_samples_per_second": 3.159,
808
+ "eval_steps_per_second": 0.79,
809
+ "num_input_tokens_seen": 4663320,
810
+ "step": 450
811
  }
812
  ],
813
  "logging_steps": 5,
814
  "max_steps": 6770,
815
+ "num_input_tokens_seen": 4663320,
816
  "num_train_epochs": 2,
817
  "save_steps": 50,
818
  "stateful_callbacks": {
 
827
  "attributes": {}
828
  }
829
  },
830
+ "total_flos": 307607720951808.0,
831
  "train_batch_size": 1,
832
  "trial_name": null,
833
  "trial_params": null